CSAT is one of the most popular satisfaction surveys. It's short, simple, and fairly intuitive.
This is why so many companies use it to gauge how satisfied their customers are with different aspects of their business.
Companies send out CSAT surveys to customers after notable customer lifecycle moments such as post-onboarding, after 6 months, prior to renewals, and following customer service interactions.
The last one - post-service conversations is a major application area of CSAT surveys.
Companies use it to understand how a customer service agent is performing by understanding how satisfactory customers are finding their interactions with an agent.
But is CSAT ratings really a true indication of agent performance?
We’re here to tell you that It’s not.
In this article, we will explore this idea in detail.
All CSAT ratings are not the same
Jack is a customer service agent at an online grocery delivery company.
He is assigned a live chat ticket where the customer is complaining about the quality of the product delivered.
He promptly attends to the customer, gathers all necessary information, and tries his best to help the customer.
The customer seeks a refund because he’s not happy with the quality of the cheese he’s received.
According to the company policy, perishable items are not eligible for return or refund.
He informs the company policy to the customer.
Jack was polite, and friendly, and answered all customer queries promptly.
The customer, however, was not pleased with the situation.
And although Jack did his best in the capacity of a service agent, the customer leaves a terrible CSAT rating.
This rating, directly attributable to Jack, makes it seem that Jack didn’t do a good job.
While the reality is quite different.
This is a common issue plaguing customer service.
Therefore, ratings received as a result of poor products and poor customer service can not be treated in the same way.
Why do you need to separate product and agent CSAT?
As seen in the example above, when companies receive CSAT ratings post customer service interactions, they all can not be treated as the same.
This is especially true when using CSAT ratings as agent performance indication.
There are several reasons for it, let’s look at some.
1. Wrong representation of service experience
Because of how CSAT surveys are framed, the negative responses are skewed towards the immediate experiences of customers.
These experiences could be from a bad customer service interaction, product problem, or operational failure.
It’s therefore important to not take the CSAT rating as an insight into a specific service interaction.
As you can see in the image above, all positive drivers are customer-service related, whereas, the negative ones are related to the product and operations.
2. Undue pressure on agents
Poor CSAT ratings could be due to a number of reasons.
So a CSAT rating post a customer service conversation should not be always attributed to the agent who handled the interaction.
In a scenario where customers leave bad CSAT ratings due to a product or operation issue, agents shouldn’t be held accountable.
Additionally, some topics always lead to bad CSAT no matter how an agent handles a customer.
This will differ depending on the type of business you're in.
For instance, for an e-Commerce company, there could be repeated product quality issues, payment-related issues, and so on.
In the image above, we should the probability of a bad CSAT rating for certain topics - no matter the quality of service.
3. Product teams can utilise their feedback better
It is important to note here that CSAT driver analysis often has a number of insights that the product teams would find useful.
However, hardly any product teams today look at CSAT or customer service conversations for insights.
Take for example a topic such as the app crashing repeatedly due to a specific reason.
If a number of users leave negative CSAT ratings because of it and if the company could dig into that specific reason, the product team could fix it leading to an overall good CX, improved CSAT rating, etc.
4. Customer service can improve with specific feedback
It is needless to say that if customer service teams knew what exactly is leading to bad CSAT ratings, they could work on those specific issues.
For example, if you could figure out the top service-related issue leading to bad CSAT, such as first response time, resolution time, etc, companies can tackle it objectively.
Rather than focusing solely on agent performance based on CSAT ratings they get.
By separating agent and product-related issues, you can easily point to the exact issues in each team.
By communicating these insights to the respective teams, you will be able to attack the true reasons for low CSAT ratings rather than blaming service agents.
This will not only lead to improved CSAT scores but also better agent productivity and ultimately elevated customer experience.