Artificial Intelligence

What’s the problem with ChatGPT in the contact center?

Ben Rigby, Talkdesk

By Ben Rigby

0 min read

Chatgpt Contact Center Ethics

Exploring the ethical concerns of ChatGPT in the contact center.

I’ve been writing feverishly lately about the opportunity ChatGPT (and other generative AI systems) presents for the contact center. At Talkdesk, we believe that the generative pre-trained transformer (GPT) large language models (LLMs) powering systems like ChatGPT are going to be the backbone of the next generation of contact centers as a service (CCaaS).

We’ve launched our first feature that harnesses these capabilities to automatically summarize customer conversations and accurately select their dispositions (e.g., requests follow-up, wants to cancel service, etc.)—effectively eliminating most of a customer service agent’s after call work. And we are actively integrating these LLMs into other core contact center use cases, as well as shaping our roadmap to accelerate the value our customers can gain from these natural language processing techniques.

But we also understand the magnitude of this new technology, and the potential it has to do harm as well as good.

A number of prominent tech leaders—including Elon Musk, Steve Wozniak, and Andrew Yang—are publicly calling for a pause in the training of AI systems more powerful than ChatGPT-4. Their general ethical concern is that runaway AI will overtake humanity’s ability to harness it for good.

With these big names weighing in on some negatives of AI systems, I thought it would be a good time to delineate some of the top ethical concerns of these new LLMs for the contact center. Any company considering using these AI models in their contact center should be aware of the potential pitfalls and consider processes that will help ensure safe and positive customer experiences.

Interested in learning more?

Connect with our team to see how Talkdesk can level-up your Call Center Software Solutions.


Five ethical concerns of ChatGPT in the contact center.

1. Transparency.

The first ethical concern with using ChatGPT and other generative AI systems in the contact center is transparency, which applies both to being clear about how an AI-driven decision was made as well as the fact that it was made by AI in the first place. Customers have the right to know if their interaction was mediated by AI. If customers believe they are talking to a human and later discover that they were talking to a chatbot, for example, it could erode trust in the company and damage customer relationships with the brand. Full disclosure to the customer is crucial when designing chatbots and virtual agents.

2. Bias.

Another ethical concern is the potential for bias in the responses of chatbots and virtual agents. ChatGPT is a machine learning model that is trained on vast amounts of text data, and there is a risk that this data could be biased in some way. If, for example, an automatic quality review favors one pattern of speech over another, this could lead to unequal treatment of agents whose backgrounds included learning that pattern of speech. In other words, this type of bias can lead to discrimination.

3. Data security and privacy.

When using ChatGPT, user interactions are logged and stored, and personal conversations could be accessible to anyone if proper security measures are not in place. As with any software system handling personally identifiable data, data flowing through the LLM needs to remain encrypted end to end and follow all of the data privacy considerations that keep customers’ data safe.

4. Truthful output.

Another ethical concern is the potential for ChatGPT to “hallucinate” or to make up answers that may not be truthful, but to convey these alternative facts in a compelling way. Of course, we don’t want to be giving misinformation to customers, especially if they’re using it to make life decisions.

5. Job displacement.

Finally, there is the concern of job displacement. As more companies turn to AI-based chatbots like ChatGPT, there is a risk that human contact center agents could lose their jobs. Companies must consider the impact that automation could have on their employees and take steps to mitigate any negative effects. In particular, contact center managers should imagine new opportunities to increase agent satisfaction by introducing tools that will automate menial tasks and help agents complete their work faster.

Chatgpt And The Contact Center Of The Future


ChatGPT and the contact center of the future

Find out how this exciting new technology will change everything from conversational AI to the role of the contact center agent.

Pause at GPT-4?

The speed of recent advances in LLMs has been astounding, and there are legitimate concerns as outlined above. The rapid development of AI without safeguards (at the very least, consideration and mitigation of the negative outcomes) could result in those negative outcomes coming to life on a much larger scale than they already have begun to do. 

The Future of Life Institute (FLI)—the author of the open letter calling for a pause—wants to ensure that AI technology is developed in a way that benefits humanity and avoids potential risks and dangers. They are concerned that the major AI labs are in an “out-of-control” arms race, and currently, there is no proper oversight ensuring they are working in an ethical manner.

We’ll be tracking these discussions carefully as they evolve over time. But our commitment to our customers and to the millions of people who interact with our software systems every day is to continuously evaluate the impact of our decisions against the ethical considerations outlined above.

So is ChatGPT ethical to use in the contact center?

In short, yes…with the appropriate precautions. For the LLM-powered features that we’ve already delivered to market and those we are developing, we at Talkdesk have already taken steps to minimize the potential ethical issues described above. We allow for transparent decision-making in the development process. We include modification by a human reviewer to remediate bias if it should exist. We maintain an absolute commitment to data privacy and security. We include anti-hallucination techniques in all our product designs to ensure no misinformation reaches our clients or their customers. And we give careful consideration to the impact of LLMs on human agents—with an eye toward imagining expanded roles for humans in the age of automation (I recently wrote about the concept of steering agents that oversee a team of bots).

While LLMs will improve the quality of service for the millions of us looking for help and support from the companies we love, it’s incumbent on the software engineers, designers, and product leaders among us to consider and address these ethical concerns as we build.

See how ChatGPT will revolutionize customer service.

Join me and Brooke Lynch, CCW Digital’s Senior Analyst, to learn how ChatGPT and generative AI is going to fundamentally change our understanding of customer service and the role of the contact center. In this conversation, I will explain why ChatGPT is so revolutionary and the impact it’s having on natural language processing. I will also share my vision of the role of the customer service agent in the era of generative AI. Watch the full webinar to see how ChatGPT is transforming customer service.

Finding this info helpful?

Learn even more about how Talkdesk can increase the quality of your Customer Experiences.



Ben Rigby, Talkdesk

Ben Rigby

SVP, Global Head of Product & Engineering, Growth at Talkdesk, unicorn contact center software as a service (CCaaS). Previously led AI at Directly: Automating customer service with AI-powered virtual agents; CEO of Using machine learning models to predict churn, retention and LTV; Software engineer at The Main Quad: acquired by Student Advantage; Engineering lead for The North Face's consumer web site for five years; CTO at SaaS startup with clients including Sam Adams, Hyundai, Old Navy, IBM, The Sierra Club and Scion.