Person typing on a laptop

Do Customers Perceive AI-Written Communications as Less Authentic?

Kim Campo| October 16, 2024

From Nike and Google to Coca-Cola and McDonald’s, major brands are incorporating artificial intelligence (AI) into their advertising campaigns. But how do consumers feel about robots generating emotionally charged marketing content? That’s the question a New York Institute of Technology professor raises in a new Journal of Business Research study.

Whereas predictive AI allows marketers to forecast consumer behavior, generative AI enables them to produce novel content, including text, images, videos, or audio. For example, a recent AI-generated Toys“R”Us commercial featured a video of the company’s founder as a boy alongside its brand mascot Geoffrey the Giraffe. While many brands have trumpeted their AI-driven campaigns as a mark of innovation, others may fail to disclose AI use, leading to ethical concerns and calls for government regulation. However, even transparent brands receive backlash, as Google experienced when viewers were offended by its “Dear Sydney” ad, in which a father uses AI to help his daughter draft a fan letter to her favorite Olympic athlete.

“AI is a new territory for brand marketers, but what we do know is that consumers highly value authentic interactions with brands,” says the study’s lead author Colleen Kirk, D.P.S., professor of marketing and management. “Although more companies are now using AI-generated content to strengthen brand engagement and attachment, no study has explored how consumers view the authenticity of textual content that was created by a robot.”

Kirk and her study co-author, Julian Givi, Ph.D., a marketing researcher and faculty member at West Virginia University, completed various experiments to see how consumers react when emotional messages are written by AI. They hypothesized that consumers would view emotionally charged AI-generated content less favorably, impacting their perception of the brand and desire to interact with it.

In one scenario, participants imagined receiving a heartfelt message from a fitness salesperson who helped them buy a new set of weights. The message stated that he was inspired by the consumer’s purchase, with some participants believing it was AI-generated and others believing that the salesman drafted it himself (control group). While the members of the control group responded favorably, those in the AI group felt that the note violated their moral principles (moral disgust). As a result, this group was also unlikely to recommend the store to others and more likely to switch brands when making future purchases. Many even gave the store poor ratings on a simulated reviews site.

Other scenarios also revealed key findings in support of the researchers’ hypothesis:

  • AI-generated factual communications (vs. emotional communications) were less likely to elicit moral disgust, reduce positive word of mouth, and diminish brand loyalty among consumers.
  • AI-generated messages in which the robot had self-autonomy (for example, an AI-generated memo signed by “The Summit Apparel AI Chatbot”) were viewed more favorably than AI-generated messages signed by a company representative.
  • When participants believed that most emotional marketing communications were written by AI, they expressed disgust. The reverse was true when they believed most communications were written by a human. Therefore, brands may benefit from promoting the human origins of their products and communications.
  • Messages simply edited (not written) by AI were less penalized for authenticity.
  • Human communicators (vs. AI) faced a greater “authenticity penalty” for copying emotional content.

In short, the findings suggest that companies must carefully consider whether and how to disclose AI-authored communications, always prioritizing authenticity in their interactions with consumers. As governments seek to increasingly regulate AI disclosure, making consumers more aware of how brands craft their messages, Kirk says marketers will want to pay close attention to the study’s findings.

“Consumers are becoming ever more skeptical of the human origin of marketing communications. Our research provides much-needed insight into how using AI to generate emotional content could negatively impact brands’ perceptions and, in turn, the consumer relationships that support their bottom lines. While AI tools offer marketers a new frontier, these professionals should bear in mind a time-tested principle: authenticity is always best,” she says.  

More News

Groups of New York Tech students

Honors College Launches

Students in New York Tech’s new Honors College will tackle real-word problems from an interdisciplinary perspective.

Portrait of Charles N. Moore

College of Arts and Sciences Welcomes New Dean

On September 1, New York Tech welcomed Charles Moore, Ph.D., as dean of the College of Arts and Sciences, where he will contribute to the university’s goal of achieving high research activity.

Three New York Tech students talking to a farmer

Guiliano Global Fellows: Glacier Saviors, Exoplanets, and More

Under the Edward Guiliano Global Fellowship Program, seven students traveled the globe, broadening their perspectives and working on transformational research projects.

Man at a desk with laptop and computer screens looking at an iPad

Staying Cybersafe

For Cybersecurity Awareness Month, Information Security and Compliance Analyst Thomas Vallario shares insight and tips to safeguard data.

Portrait of Lloyd Jura

From the Corporate World to Higher Ed

After serving on the College of Engineering and Computing Sciences’ advisory board and teaching a course in risk management, New York Tech-Vancouver Assistant Professor Lloyd Jura, Ph.D., found a new career path: educator.

OMED sign

NYITCOM Presents at OMED 2024

College of Osteopathic Medicine students, faculty, and alumni shared scientific findings in San Antonio for this year’s Osteopathic Medical Education Conference.