CustomerSure

In this Guide…

CX leaders are bombarded with AI solutions — but how many actually deliver real value?

Every CX vendor claims to be ‘AI-powered’ — but how much of it is just marketing fluff?

In this guide, we’ll cut through the noise. First, we’ll examine the AI marketplace (including new developments like DeepSeek) and separate strengths from hype. Then, we’ll show how top CX teams are actually using AI to save time, improve experiences, and prove the value of CX. Along the way, you’ll develop a deeper understanding of how modern AI really works—so you can make smarter decisions anywhere AI touches your role.

AI in CX: Where’s the current value?

This article is going to cover a lot of ground, and by the end of it you’ll be able to make informed decisions about how to use AI in your area of responsibility.

As we already discussed, there are a lot of grand claims made about the future of AI, but we are going to focus on what’s working well in the here-and-now. What the pragmatic teams who we work with are seeing value in today. If grander claims, like multi-channel ai-based personalisation become commonly operationalised, we’ll update this guide.

You can of course skip to the good bit, but you’ll be more informed if you take a minute to catch up on a bit of history, and understand the strengths and weaknesses of the tech that underpins modern AI, before we look at specific applications of this tech.

How did we get here?

AI/ML (Machine Learning) has been around a long time. But practical business uses started to become more commonplace from around 2015 to 2020 when firms like Google and Amazon made available high quality, inexpensive NLP (Natural Language Processing) tools which could detect customer sentiment in feedback, and pull out key topics and themes.

These NLP tools were good, although not as impressive as the ones we enjoy today, and were quickly integrated into CX insight platforms like CustomerSure, becoming ‘table stakes’ for serious feedback analysis.

GPT and ChatGPT

However, a watershed moment for AI came with the public release of ChatGPT in November 2022. Attracting 100 million users in just two months, ChatGPT is arguably the fastest-adopted piece of technology ever.

But ChatGPT is just a chatbot ‘wrapper’ around OpenAI’s core ‘GPT’ model. This was initially GPT 3.5, but the more recent GPT 4, GPT-4o and o1 rapidly added new features and improved GPT’s already-impressive capabilities.

Most people know GPT for text generation: Chatbots, reports, code, and web content. But ironically, even ChatGPT itself doesn’t claim this is its best use case, instead hedging that it’s best used for a range of tasks, like tutoring, editing human-written text, and creative brainstorming.

One other thing that GPT is good at? customer feedback analysis, including sentiment analysis and topic detection.

GPT competitors

GPT is just one example of a Large Language Model (LLM). Although ChatGPT was the line in the sand which resulted in mass-awareness of the power of AI, it’s far from the only LLM in town.

Many, many other LLMs exist, from commercial offerings like Google’s Gemini, Anthropic’s Claude, or freely-available products like Facebook’s Llama. These LLMs are all comparably powerful and can be used to accomplish similar tasks.

Open/Free LLMs (DeepSeek and Llama)

In 2023, an internal Google memo, “We have no moat” famously leaked. The thrust of the memo is that without a ‘killer’ product wrapped around the core LLM model, any company developing its own AI technology has a rocky commercial future, because other research groups are giving away equally-powerful tools for free.

This memo turned out to be spot-on when, over Christmas 2024, DeepSeek made their V3 LLM freely available.

DeepSeek outperforms GPT on a number of benchmarks and does so at a significantly lower processing cost, meaning it’s now feasible for any company to operate their own GPT-4o-equivalent model, on their own private servers, and do so more cost-effectively than paying OpenAI.

This means CX teams now have real choices: They can run their own completely private LLMs (subject to IT resources and infrastructure), reducing data privacy issues and simplifying software subscriptions.

For some teams, this could mean significant cost savings and fewer data governance headaches compared to commercial LLMs. A CX platform (that’s us!) should be equally happy working with a team’s private LLM as they are using best-in-class commercial solutions like GPT-4o and Claude

Like other breakthroughs which have powered historic economic transformations, like coal, electricity and computer networks, AI ‘engines’ are now a commodity: Cheaply and widely available. But just like these historical commodities, it’s possible to build wildly-different products using the same building blocks.

Image showing two cars, the Jaguar X-Type and Ford Mondeo. Same platform, but one is visibly styled as an executive car, the Mondeo is more mass-market. This is a visual analogy for how identical LLM cores can result in vastly different products

How do LLMs work for feedback analysis?

If no AI provider has a ‘secret sauce’ which gives them a competitive edge, then how do you choose? And how do you get the best results? It first helps to get a good understanding of how to get good results from a LLM.

(We’ll skate over the details of exactly how LLMs perform their magic, but keen readers might enjoy this blog post which explains a lot of the key concepts without reaching for too much maths or jargon).

If you ask ChatGPT to analyse customer feedback, it will do an adequate job, probably better than what off-the-shelf tools in 2017 would manage. But of course, ‘adequate’ isn’t the goal.

When platforms like CustomerSure use a LLM for customer feedback analysis, they add a number of ‘layers’ on top of the raw LLM to produce higher-quality results. In short, these layers are:

1. Fine-tuning

LLMs provide better responses when they have more context on a given domain (i.e. your business). So a key step is providing the LLM with a lot of background and context on who you are, how you make your money, how you’re structured, and what your existing customers say about you.

2. Prompt Engineering

The ‘prompt’ to a LLM is the text that you send to the AI ‘core’ which causes it to ‘think’ and send you a response. For example,

Analyse the following customer feedback:

The follow-up communication was frequent, but the resolution itself wasn’t fully satisfactory

This basic prompt works as expected. But how useful is it?

A screenshot illustrating the above prompt in a LLM, Claude

To get more accurate, more actionable results, you need a high-quality prompt. Anthropic run a great interactive course in prompt engineering which will give you an insight into what goes into writing a good prompt. Suffice to say, the more effort and experience which goes into the prompt, the higher-quality the analysis will be.

3. Recursive Critique

Finally, it’s important to always be aware that although they’re intimidatingly powerful, LLMs will tell lies. They will simply occasionally say things which aren’t true, so this inevitably applies to feedback analysis carried out by a LLM.

There’s no guaranteed defence against this, it’s a problem that’s baked-in to how the technology works. But it’s a problem that can be mitigated through techniques like recursive critique, where instead of accepting the first answer given by a LLM, you subject that answer to further scrutiny before you accept it.

Think of it like a writer revising an article. Instead of publishing the first draft, they re-read it, check for errors, and refine weak points before publishing.

In Summary

LLMs work by receiving text (the ‘prompt’), and by turning that text into numbers and doing some very fancy maths, deciding which text to send back in response to that prompt.

An ‘out-the-box’ LLM will do an acceptable job of classifying customer feedback, but a customer experience management platform can improve the quality and accuracy of that analysis through a combination of fine tuning, prompt engineering and recursive critique.

  • A basic, un-tuned LLM might see “Long wait times for support” as a generic “customer service” issue.
  • A fine-tuned model understands this as a “queue management” issue
  • With recursive critique, slightly ambiguous feedback, which could — for example — be classified as “queue management” or “tech” if the long wait times were caused by a slow website, will end up classified correctly.

Finally, analysis accuracy is only half the story. GPT-4o might get it right 95% of the time, versus a ‘trained’ LLM getting it right 99.9% of the time, but neither will add to your bottom line unless it’s easy to reach insights from that analysis, and put those insights into the hands of teams empowered to improve things for customers.

How to get the most out of LLMs/GenAI in Customer Experience

Nothing in life comes for free. LLMs are shockingly flexible, powerful tools, great at feedback analysis… But to make best use of them, it’s important to understand where they fall short.

We’ll also cover a few areas where the public perception of LLMs (often formed from consumer interactions with ChatGPT in 2022) has not kept pace with the technical reality in 2025.

Are LLMs biased when they analyse feedback?

Yes. LLMs are trained on vast amounts of data written by humans, and then further fine-tuned by humans to give ‘desirable’ outputs based on that data.

The outputs of any LLM are biased based on the biases of these inputs.

How this affects your use of a LLM depends on how you’re using it.

  • If you’re, for example, using AI to do any sort of customer profiling, based on demographic categories, you should be very wary of potential biases. At the very least you should intentionally anti-bias your prompts (or more likely, ensure your supplier has done this), and ensure that humans are able to easily override this profiling in any cases where bias has crept in.
  • If you’re using AI to analyse customer feedback, the risks of bias are less severe, but they still exist.

    The two key biases to avoid are positivity bias and cultural bias: In short, LLMs may be too optimistic about borderline cases, and may misunderstand feedback which would be understood differently in different cultural contexts (think US English vs UK English). ChatGPT itself gives these three examples of customer feedback it could misinterpret, based on the cultural context:

“Not bad at all.”

🇬🇧: “Actually quite good.” (Positive)

🇺🇸: More neutral or even slightly negative, implying it’s just acceptable. (Neutral/Negative)

“Quite good, but could be better.”

🇬🇧: “Moderately good, but not great.” (Neutral/Negative)

🇺🇸: Sounds more like “very good.” (Neutral/Positive)

“Thanks for sorting that out. Cheers.”

🇬🇧: “Cheers” is fairly routine in the UK. (Weak positive)

🇺🇸: “Cheers” is often associated with celebration or strong approval. (Strong positive)

As you can see, these biases tend to be edge cases, but they do exist.

Can You Trust AI to Give the Same Answer Every Time?

Not necessarily. LLMs are probability-based, which means they can give different outputs for the same input. And in ultra-rare edge cases, those outputs won’t even necessarily make sense. However, you need to bear in mind two key considerations:

  • For ‘narrow’ tasks like feedback classification, the chances of wildly-different outputs are very low (compared to a prompt which asks for ‘creative’ work). Although it’s theoretically possible, you’re unlikely to see it often in real-world use.
  • The cases which do result in non-repeatable results are almost always borderline. If you asked five different humans to classify them, you might receive three different responses. It’s a case where there is no ‘correct’ answer, so it’s natural that the LLM responds in a similar way.

As with everything we’ve discussed so far, once you’re aware of how the underlying tools work, you (or your CX platform partner) understand how to use techniques like prompt engineering and recursive critique to get the best results.

AI & Energy Use: Overblown Concern or Real Problem?

The honest answer is, it’s complicated.

The worst-case scenarios which circulate on social media about LLM energy consumption are rarely entirely accurate. Whilst training models, and servicing requests does consume a large amount of power, efficiency is rapidly increasing.

Consider that both OpenAI and Anthropic have approximately halved their per-token pricing between 2023 and 2025 because it costs them less energy to serve each request.

Or consider how much more cost-effective DeepSeek is than both. This 10x decrease in cost-per-token is largely due to a huge increase in power efficiency.

Or read this hugely in-depth research into the power consumption of AI which reports that modern Nvidia GPUs are 25x more power-efficient than the previous generation.

None of this means that LLMs are so absurdly energy-efficient to run that we shouldn’t care about their power consumption. It just means that the energy efficiency story in 2025 is vastly different to how it was reported in 2022.

Combined with the fact that much of this new power demand is being met by investments from datacentre builders into renewables (see the Bloomberg report for more on this), it means that although every company needs to carefully consider their carbon footprint, AI isn’t uniquely bad: It’s simply another energy cost that needs to be budgeted for, minimised, and offset if necessary.

Are there privacy concerns using AI to analyse feedback?

If you’re doing it right, no.

It’s good that we live in a world which is more aware of data security than ever, but this particular concern feels rooted in consumer experimentation with ChatGPT, rather than business-grade use of LLM tools.

If a consumer uses the web version of ChatGPT, they by-default opt in to having their data used to improve future GPTs. So under no circumstances should you experiment with using ChatGPT like this to analyse customer data.

However for commercial GPT-4o use, data is not used for training. Google do not use your data to train Gemini. Anthropic (mainly) do not use your data to train Claude.

This doesn’t mean we should become careless about personal data. It’s important that we work within the framework established by GDPR to send surveys and process customer data, which means:

  • Don’t send more data than is necessary to your CX technology partner. Enough PII to identify and segment customers is fine and necessary, but you shouldn’t be sharing their mother’s maiden name and the place they were born.
  • Ensure your tech partner can comply with their GDPR obligations, for example respecting your data rentention period, respecting laws around sending email, and supporting individuals’ right to data rectification.

Any other strengths or weaknesses?

No AI system — from classical NLP systems to modern LLMs is flawless when analysing highly-nuanced feedback, especially when sarcasm is thrown into the mix. It’s fair to say that many humans struggle with this as well. LLMs outperform the competition here due to their greater flexibility and contextual understanding, but they can’t operate flawlessly.

Of course, as with the issues around bias and non-repeatability of results, a ‘trained’ LLM can outperform an ‘out the box’ LLM to such an extent that this isn’t an issue in day-to-day use.

What’s most important to CX teams when selecting an AI/ML tool?

Technology doesn’t make or break a CX project — execution does.

Beware vendors selling a “magic AI edge.” The history of the entire software industry to date shows that nobody has a technical edge, or at least not one that lasts for long. What matters to customers (and therefore your bottom line) is how well you apply insights, empower teams, and act on customer feedback.

Over the space of three years we’ve seen OpenAI move from being an unassailable market leader, to closely-matched by Anthropic, to seeing DeepSeek giving away a tool which can perform the same tasks as either at a tenth of the cost.

A CX platform which utilises any of these products, or their own “secret sauce” will do an exemplary job of classifying customer feedback by sentiment and topic. But on its own, that classification isn’t going to delight your customers, and it isn’t going to contribute to your bottom line.

What matters is the insights you derive from that feedback, how you act on those insights, and how you handle at-risk customers who have urgent needs you need to meet.

A fully-featured CX platform, like CustomerSure — or any of our competitors — is able to handle the technical side of feedback analysis. But if you’d benefit from a genuine CX partner, who will help you map journeys, work with stakeholders across the business, and gently challenge you with best practice, we can do that too.

How CX teams are using AI

Finally, now you know a lot more about how modern AI works, its pros and cons, and how to mitigate those cons, let’s look at what other CX teams are doing well. All these examples are based on the real-world experiences of the insight/CX teams we work with.

VoC insights across multiple channels.

There’s no longer any excuse for customer data to remain silo’ed.

We work with teams who use CustomerSure to ingest data from touchpoints and channels across the whole business, slice that data in a way which reflects both customer segmentation and operational structure, and accurately analyse that data using powerful AI tools.

This structured analysis lets those teams demonstrate their value to the board by delivering with a whole-business view of satisfaction, and empower product/service owners with clear, relevant insights that they can act on to make customers’ lives better.

Vulnerable customer detection

It’s not just Consumer Duty: Many firms are under pressure to demonstrate that they identify and act on vulnerable customers.

Rather than simple topic analysis, some of the best teams we work with also use AI to spot customers who may have characteristics of vulnerability, even if those customers don’t realise it themselves.

Of course, as we discussed, it would be wrong to let a LLM profile these customers without any right of rectification, but this does provide a valuable ‘amber flag’ that teams use to prioritise customers for intervention.

Customer Feedback Analysis

We’ve focused heavily on feedback analysis in this guide, but this is because it’s one of the most obvious use-cases for LLMs: tools which by their very nature turn text into numbers.

Teams we work with use our AI tools to scan feedback for overall, and per-topic customer sentiment. Once the feedback is enriched with this topic and sentiment data, teams can automatically spot negative feedback, and instantly send it to the right person in the business to fix things

Screenshot showing the topic "Application process" with a positive sentiment applied to some customer feedback

Translation and Speech-to-Text

Translation and Speech-to-Text are both examples of how LLMs can act as force-multipliers. Even when they serve highly-diverse customer bases (either globally, or in a multicultural region), not all CX teams are as well funded as they deserve to be.

Whilst still lagging behind true bilingual human translation, LLMs, and more specialised ML translators such as DeepL are inexpensive and deliver results which are often a lot better than ‘good enough’.

You wouldn’t use these tools to — for example — re-write critical safety or compliance text from English into another language, but they are more than capable of providing an accurate translation of what — for example — a Polish tenant is telling you in a customer survey, and understanding if it needs an urgent response.

Action Planning

It can be challenging to engage people in operational roles with CX initiatives. Most teams will be familiar with colleagues who truly care about customer satisfaction but don’t have the time to read every single customer comment to work out how to prioritise things.

AI can produce short action plans to engage senior leaders in other areas of the business.

Screenshot showing the topic "Application process" with a positive sentiment applied to some customer feedback

With good fine-tuning (i.e. when the AI understands your organisational structure well), these action plans can condense thousands of customer comments to a few actionable bullet points. Time saved on analysis gets spent on delighting customers.

Rapid prototyping

Returning to the list of tasks that ChatGPT considers itself good at, it’s no surprise that we see CX teams using LLMs as tutors and coaches, with astonishingly good results.

When you take motivated, smart people (and most people in CX are nothing if not motivated and smart), and give them access to the sum of human knowledge through an interface that acts like a patient teacher, amazing things happen.

We’ve watched teams teach themselves basic coding and data analysis skills, learning to build in-house tools that integrate with CustomerSure that would ordinarily have a lead-time of months if they were routed through a centralised IT department.

This makes these CX teams more agile, and gives them greater job satisfaction: Rather than being held hostage by technology, they’re learning to use the tools they have to build better outcomes for customers.

Summary: How to (really) improve CX using AI

  • Build a good ‘basic’ understanding of how LLMs work, so you understand how to get the best out of them.
  • Get comfortable using an ‘out-the-box’ LLM for ad-hoc tasks, such as paired learning and brainstorming. This should be a paid, private commercial version of GPT-4o or Claude, or something like a private version of DeepSeek that your organisation hosts.
  • For more structured, ongoing tasks such as feedback analysis and action planning, partner with a firm like CustomerSure. Our team will work alongside you to ensure that CX makes a measurable difference to your bottom line.
Chris Stainthorpe
Chris Stainthorpe

Chris is a member of CustomerSure’s founding team. Since 2010, he has worked with our clients to help them put VoC programmes in place that respect the customer and deliver measurable results. The best part of his job is learning something new from every new client.

Take the first step…

Connect with a CX expert who’ll help determine your current VoC programme maturity level and provide a 3-step action plan to improve.