When humanlike chatbots miss the mark in customer service interactions
Listen to this article
I’ve got to confess, one of my pet hates is chatbots—especially ones that sound humanlike. For that reason, this study appealed to me.
A new paper in the Journal of Marketing examines the use of chatbots in customer-service roles and finds that when customers are angry, humanlike chatbots can negatively impact customer satisfaction, overall firm evaluations, and subsequent purchase intentions.
Chatbots are increasingly replacing human customer-service agents on companies’ websites, social media pages, and messaging services. Designed to mimic humans, these bots often have human names (e.g., Amazon’s Alexa), humanlike appearances (e.g., avatars), and the capability to converse like humans. The assumption is that having humanlike qualities makes chatbots more effective in customer service roles. However, this study suggests that this is not always the case.
The research team found that when customers are angry, deploying humanlike chatbots can negatively impact customer satisfaction, overall firm evaluation, and subsequent purchase intentions. Why? Because humanlike chatbots raise unrealistic expectations of how helpful they will be.
The researchers conducted five experiments to better understand how humanlike chatbots impact customer service.
Study 1 analyzed nearly 35,000 chat sessions between an international mobile telecommunications company’s chatbot and its customers. Results show that when a customer was angry, the humanlike appearance of the chatbot had a negative effect on the customer’s satisfaction.
Study 2 was a series of mock customer-service scenarios and chats where 201 participants were either neutral or angry and the chatbot was either humanlike or non-humanlike. Again, angry customers displayed lower overall satisfaction when the chatbot was humanlike than when it was not.
Study 3 demonstrated that the negative effect extends to overall company evaluations, but not when the chatbot effectively resolves the problem (i.e., meets expectations). More than 400 angry participants engaged in a simulated chat with a humanlike or non-humanlike chatbot and their problems were either effectively resolved or not during the interactions. As expected, when problems were not effectively resolved, participants reported lower evaluations of the company when they interacted with a humanlike chatbot compared to a non-humanlike one. Yet, when their problems were effectively resolved, the company evaluations were higher, with no difference based on the type of chatbot.
Study 4 was an experiment with 192 participants that provided evidence that this negative effect is driven by the increased expectations of the humanlike chatbot. People expect humanlike chatbots to be able to perform better than non-humanlike ones; but those expectations are not met, leading to reduced purchase intentions.
Study 5 shows that explicitly lowering customer’s expectations of the humanlike chatbot prior to the chat reduces the negative response of angry customers to humanlike chatbots. When people no longer had unrealistic expectations of how helpful the humanlike chatbot would be, angry customers no longer penalized them with negative ratings.
What the researchers say: “Our findings provide a clear roadmap for how best to deploy chatbots when dealing with hostile, angry or complaining customers. It is important for marketers to carefully design chatbots and consider the context in which they are used, particularly when it comes to handling customer complaints or resolving problems.”
Firms should attempt to gauge whether a customer is angry before they enter the chat (e.g., via natural language processing), they say, and then deploy the most effective (either humanlike or not humanlike) chatbot. If the customer is not angry, assign a humanlike chatbot; but if the customer is angry, assign a non-humanlike chatbot.
If this sophisticated strategy is not technically feasible, companies could assign non-humanlike chatbots in customer service situations where customers tend to be angry, such as complaint centers. Or companies could downplay the capabilities of humanlike chatbots (e.g., Slack’s chatbot introduces itself by saying “I try to be helpful. But I’m still just a bot. Sorry!” or “I am not a human. Just a bot, a simple bot, with only a few tricks up my metaphorical sleeve!”). These strategies should help avoid or mitigate the lower customer satisfaction, overall firm evaluation, and subsequent purchase intentions reported by angry customers towards humanlike chatbots.
So, what? Or, of course, why not dispense with the bots altogether? Many studies have shown that the rise in the level of depression, burnout and other mental illnesses can be traced to the increasing dehumanization of our workplaces and our society.
There is no reason whatsoever we should allow this dehumanization to continue as it is profoundly against our design specs and that which is against our design specs causes us to be stressed, depressed and anxious.
Bots are there simply to make money for the firms that deploy them. No technology should be allowed unless it can be proved that it increases the overall rate of human happiness.
Join the discussion
More from this issue of TR
You might be interested in
Back to Today's ResearchJoin our tribe
Subscribe to Dr. Bob Murray’s Today’s Research, a free weekly roundup of the latest research in a wide range of scientific disciplines. Explore leadership, strategy, culture, business and social trends, and executive health.