News — OLD WESTBURY, N.Y. — From Nike and Google to Coca-Cola and McDonald’s, major brands are incorporating artificial intelligence (AI) into their advertising campaigns. But how do consumers feel about robots generating emotionally charged marketing content? That’s the question a New York Institute of Technology professor raises in a new  study.

Whereas predictive AI allows marketers to forecast consumer behavior, generative AI enables them to produce novel content, including text, images, videos, or audio. For example, a recent AI-generated Toys“R”Us commercial featured video content of the company’s founder as a boy alongside its brand mascot Geoffrey the Giraffe. While many brands have trumpeted their AI-driven campaigns as a mark of innovation, others may fail to disclose AI use, leading to ethical concerns and calls for government regulation. However, even transparent brands receive backlash, as Google experienced when viewers were offended by its “” ad, in which a father uses AI to help his daughter draft a fan letter to her favorite Olympic athlete.

“AI is a new territory for brand marketers, but what we do know is that consumers highly value authentic interactions with brands,” says the study’s lead author , professor of marketing and management at . “Although more companies are now using AI-generated content to strengthen brand engagement and attachment, no study has explored how consumers view the authenticity of textual content that was created by a robot.”

Kirk and her study co-author, , a marketing researcher and faculty member at West Virginia University, completed various experiments to see how consumers react when emotional messages are written by AI. They hypothesized that consumers would view emotionally charged AI-generated content less favorably, impacting their perception of the brand and desire to interact with it.

In one scenario, participants imagined receiving a heartfelt message from a fitness salesperson who helped them buy a new set of weights. The message stated that he was inspired by the consumer’s purchase, with some participants believing that it was AI-generated and others believing that the salesman drafted it himself (control group). While the members of the control group responded favorably, those in the AI group felt that the note violated their moral principles (moral disgust). As a result, they were also unlikely to recommend the store to others and more likely to switch brands when making future purchases. Many even gave the store poor ratings on a simulated reviews site.

Other scenarios also revealed key findings in support of the researchers’ hypothesis:

  • AI-generated factual communications (vs. emotional communications) were more likely to elicit moral disgust, thus reducing positive word of mouth, and diminishing brand loyalty among consumers. 
  • AI-generated messages in which the robot had self-autonomy (for example, an AI-generated memo signed by “The Summit Apparel AI Chatbot”) were viewed more favorably than AI-generated messages signed by a company representative.
  • When participants believed that most emotional marketing communications were written by AI, they expressed disgust. The reverse was true when they believed most communications were written by a human. Therefore, brands may benefit from promoting the human origins of their products and communications.
  • Messages simply edited, but not written, by AI were less penalized for authenticity.
  • Human communicators (vs. AI) faced a greater “authenticity penalty” for copying emotional content.

In short, the findings suggest that companies must carefully consider whether and how to disclose AI-authored communications, always prioritizing authenticity in their interactions with consumers. As governments seek to increasingly regulate AI disclosure, making consumers more aware of how brands craft their messages, Kirk says marketers will want to pay close attention to the study’s findings.

“Consumers are becoming ever more skeptical of the human origin of marketing communications. Our research provides much-needed insight into how using AI to generate emotional content could negatively impact brands’ perceptions and, in turn, the consumer relationships that support their bottom lines," she says. "While AI tools offer marketers a new frontier, these professionals should bear in mind a time-tested principle: authenticity is always best."

 

About New York Institute of Technology

New York Institute of Technology's six schools and colleges offer undergraduate, graduate, doctoral, and other professional degree programs in in-demand disciplines including computer science, data science, and cybersecurity; biology, health professions, and medicine; architecture and design; engineering; IT and digital technologies; management; and energy and sustainability. A nonprofit, independent, private, and nonsectarian institute of higher education founded in 1955, it welcomes nearly 8,000 students worldwide. The university has campuses in New York City and Long Island, New York; Jonesboro, Arkansas; and Vancouver, British Columbia, as well as programs around the world. More than 116,000 alumni are part of an engaged network of physicians, architects, scientists, engineers, business leaders, digital artists, and healthcare professionals. Together, the university's community of doers, makers, healers, and innovators empowers graduates to change the world, solve 21st-century challenges, and reinvent the future. For more information, visit .

MEDIA CONTACT
Register for reporter access to contact details
RELEVANT EXPERTS
CITATIONS