Actionable insights from qualitative feedback performance comments with text analytics and AI
Every year, your firm collects thousands of words in performance reviews and feedback. Often these words have been agonised over, rewritten and debated. So much labour - financial and emotional - have gone into producing these words.
How much of that rich, qualitative data is actually being used to its full potential? Hopefully it’s led to a great review conversation and encouraged deep reflection, but beyond that most of it would sit untouched, because analysing it is time-consuming, overwhelming, and expensive.
Natural Language Processing (or NLP for short) was born in the 1940s to solve the problem of language translation at scale. It’s since grown into many other applications - from smart email filters that detect malicious emails to banking tools that flag fraudulent transactions. We already trust NLP to guide attention and support decisions.
In brief, NLP involves using computers to 'read' text and
Large Language Models (the engine behind AI tools like ChatGPT, CoPilot, Gemini, Claude, DeepSeek) perform Natural Language Processing. These computer models are trained on billions of examples of text and can recognise words and patterns in language that indicate meaning and sentiment. For example, they can detect that "The workload is over overwhelming" expresses negative sentiment about the topic of "workload."
It’s like having a team of assistants who can read through all your feedback, highlight the key takeaways, and tell you what’s really going on beneath the surface - except they do it in seconds and they are far more consistent than a team of human assistants (human judgement is inconsistent and noisy).
I used the above prompt in Microsoft CoPilot to analyse the three sample sentences in the the first column in the table below. The idea is to test how AI might understand qualitative comments that might be expressed at your firm. You should try it with your own examples.
By turning feedback, appraisals and objectives into actionable insights, topic and sentiment analysis helps you make smarter decisions about talent development, team structure, and organisational culture. For example:
The potential is clear. The challenge is how to do this across large amounts of data (without losing context) in a quick and easy way.
At Performance Leader, we're integrating AI to increase high quality feedback, support meaningful conversations and drive insights from data. The following approaches have show great promise in internal research and client testing.
The Weighted Net Sentiment Index is a single metric that measures the overall sentiment in text data like feedback or reviews. It's calculated as the difference between positive and negative sentiment percentages, with a weight applied to the negative scores.
We tested multiple approaches in our research and like this one because it
Below are some examples of what the score would generate and what it could mean - the shaded columns capture the index score and what it could suggest based on the positive, neutral and average sentiment scores.
The charts below show how we display this sentiment index over your data and highlight the difference between sentiment and WNSI. In our research, admins found the chart on the right more useful because
The Weighted Net Sentiment Index can also be used to explore sentiment across topics across different roles, departments or projects. The chart below extract skills and organisation names, and allows admins to apply that view over different departments or roles, so a leader can answer questions like
Key Topic - Descriptor Pairs are a way to analyse and summarise text data by identifying the main topics (key topics) and the specific words or phrases (descriptors) that describe or relate to those topics. It consists of a broad subject or theme (like "workload," "communication," or "career growth") and the specific words that describe it (such as "too high," "unclear expectations," or "lack of opportunities").
These are far more powerful than simple word clouds, because they add valuable context and thus insight into why certain topics are being discussed - for example, moving beyond simply noting "workload" to understanding that "workload is too high."
Below is a preview in Performance Leader.
Key topic - descriptor pairs are valuable tools in both performance reviews and client feedback contexts. In performance evaluations, they effectively identify strengths and areas needing improvement, such as noting "team collaboration" as "excellent during projects." Similarly, when analysing client feedback, they help pinpoint specific satisfaction drivers and areas of concern, like identifying "response time" as an aspect that "needs improvement."
Natural Language Processing and text analysis tools represent a significant leap forward in how organisations can utilise their qualitative data. By implementing these AI-powered solutions, firms can quickly extract actionable insights from thousands of words of feedback and drive meaningful improvements in performance management and organisational development.
The combination of Weighted Net Sentiment Index and Key Topic-Descriptor Pairs provides a comprehensive framework for understanding not just what people are saying, but how they feel about it. This deeper level of analysis enables leaders to:
It's never been easier to experiment with Natural Language Processing and sentiment analysis. Run text from different contexts in your favourite AI model and see what you can learn. Which types of comments stand out? What is not obvious, and how can you dive deeper?
As we continue to develop and refine these tools at Performance Leader, our goal remains clear: to support more frequent, easier and better feedback interactions between people, and harness the full potential of your feedback data to create more effective, engaged, and successful teams.