Text Analytics as used in Contact Centers


     
Evolution of Unstructured Data in Contact Centers
Having spent more than 15 years in contact centers that been the focal point of customer interactions of all kind – signing up for a credit card, subscribing for an energy connection to getting your computer issues solved through phone or an online channel to requesting for cancellation from different subscription businesses – one thing I see common amongst the half a dozen organizations that I have worked for. All of them handle in excess of 10 million transactions per month and there is a huge opportunity to understand customer behavior during different stages of the product lifecycle. All these organizations handle structured data around these interactions extremely well. There are many meaningful reports that are churned out of this data from different systems –Switch, Ticketing System, Surveys, Order Management System etc..
Around 2004-05, most US clients started looking at Philippines as an option compared to Indian cities. The fact that the Americans ruled the South East Asian nation had definitely had a role to play in its present culture which is so much “Follow the Yankees”.  A little later, around 2007, webchat started playing a huge role in the contact mix of CARE interactions. A lot of Fortune 500 organizations started their digitization strategy with chat and then virtual avatar aka intelligent self-service strategies.  This ensured the availability of the text transcript [chat] in addition to all the data around it – like which page did the customer come from, what keyword did he use, what IP address did he use, which region is that IP address assigned to etc…This started the revolution of the usage of unstructured data analysis in contact centers.
Text Analytics defined…
Text Mining refers to the process of deriving high-quality information from text. This is achieved by transposing words and phrases in unstructured data into numerical values which can then be linked with structured data in a database and analyzed with traditional data mining techniques. The application of text mining techniques to solve business problems is called text analytics. A lot of websites will refer to both “text mining” and “text analytics” interchangeably.
Text Mining Architecture
The text mining process can be based on either linguistic or statistical or a combination of both approaches. Gartner breaks the entire ecosystem into 3 stages –
1.       Data Collection [Text acquisition and preparation]
2.       Data Processing [processing and analysis] and
3.       Data Representation [end output]




A High Level Technical Architecture of a Text Analytics System is provided in the framework below
Source: Gartner (August 2014)

More details of the activities and outputs in each stage are provided in the table below.






In this paper, we will discuss the data processing aspect and less on the 1 -data collection and 3- data representation side.
2.      Data Processing
Machine Learning for Unstructured Data
Unstructured data offers terabytes of information. That presents a huge opportunity as well as a big challenge to synthesize meaningful information. Machine learning provides a huge opportunity to synthesize the volume of data that is available and is one of the most common approaches to deal with large volumes of data. Machine learning is a scientific discipline that explores the construction and study of algorithms that can learn from data. Such algorithms operate by building a model based on inputs and using that to make predictions or decisions, rather than following only explicitly programmed instructions. Machine Learning can be considered a subfield of Computer Science and Statistics. In 1959, Arthur Samuel defined machine learning as a "Field of study that gives computers the ability to learn without being explicitly programmed"
Most researchers suggest that Linguistic approaches are best used for text analysis instead of machine learning techniques. Here is an example of how a leading text mining software categorizes some of the text available in a survey based on the standard library.
“Get rid of the piece of crap automated system you operate and give me an option to go to a knowledgeable human directly.” – categorized as an Agent Knowledge issue by the text mining software whereas this is a complaint about the IVR
“The ease of sending & receiving money without worrying about my details falling into the wrong hands” - – categorized as an Agent “Easy to Understand” by the text mining software whereas this is more about the process of transferring money
Linguistic Approaches as an alternative for Machine Learning Techniques
Natural language text is not a medium readily understood by computer systems, in contrast to the neatly arrange rows and columns in a database. This is the primary reason that text analytics has had such a long gestation before it could be usefully employed in a business arena. Our experience suggests it is best to use deep linguistic analysis to generate a stable data dictionary with high coverage and accuracy and then use machine learning to automate these. In the first step, which is to identify words A combination of these approaches provide the best possible when the data volumes are high.  Coverage and Accuracy around 70% is a tall order in such cases whereas business expects coverage and accuracy of 90%+.
The data preparation aspect brings the need for text tagging. Text Tagging is done in 2 different ways as of now
1.       Human Tagging : This is non-scalable and subject to differences between different resources
2.       Auto Tagging : Uses an automated system based on unsupervised machine learning techniques
How do you combine the best of human tagging and auto tagging to generate a cost-effective solution?
1.       Use the keyword extraction feature of the text analytics tool to look at the keywords or phrases that show the # of occurrences
2.       Unsupervised category-document correlation generates numerical scores for each document relative to a set of interesting topics.
3.       Have a human tagger with process knowledge of the interactions use these keywords and phrases to build the initial set of data dictionaries
4.       Regression with supervised machine learning generates numerical probability that a document belongs to one of many 'classes' established from ground-truth.





CONTEXT is KING
While these are examples of incorrect text categorization, it sends a larger message that CONTEXT is more important in Unstructured Data Analytics. Definition of Context is a challenge that every linguistic analyst is set to conquer. A good linguistic approach uses the following approaches
1.       A data dictionary that works with keywords does not provide you with an accurate insight as languages are more complex e.g. social security number being picked up for security topics. Look for phrases. The expression "kick the bucket" may not make sense to someone unfamiliar with American idioms.
2.       Relation Extraction when sentences follow subject, action and object order instead of a bag of words approach.
3.       Distance is a key factor in dialogues as most context in a response  is appropriate only in the immediate distance instead of the entire length of the chat text
4.       Modifiers and Negations– It is important to use modifiers and Negations. Adverbs and Adjectives are used to qualify a verb and/or noun. There is a big difference between “good”, “very good” and “not so good”
5.       Sentiment Analysis – While providing a sentiment for the keyword is something most text analysis engines can achieve, providing a sentiment score for each document based on the modifier and the negation for the whole document
6.        Visualization and Search Access – It is important to allow the user to visualize the analysis as well as to provide search functionality for the business user to understand the context better
How is text mining combined with predictive analytics effective in Customer Experience Management?
Text Mining of the verbatim in surveys and the chat transcript allows the organization to identify contextual variables that add more details to the existing structured data set for the transaction. In addition, there is an opportunity to provide a sentiment score for the interaction. This score and the other variables allow prediction of the NPS for 100% of the interactions – a huge improvement from the 5-20% response rate that is typically available. A single view of all these variables [structured, unstructured and predicted] for the leadership and alerts for poor predicted NPS allows the Organization to close the loop with not just the detractors who responded to the survey but also those who in other cases would have just switched.


Build versus Buy
Like every technology solution, there is an option for every user to buy, build or partner with a provider for re-use.  Some of the criteria that determine this are as follows
1.      Customization needed for different engagements and different data sourcesIf you are trying to set the data dictionaries once and have minimal changes and you do not need to constantly add new data sources
2.      Integration to other systems and Support needed – If you are looking to integrate this to your social media listening tool or your customer experience management tool in the contact center or a Visual analytics tool for multiple users
3.      Skillset available in the organization
4.      Cost

Text Analytics Providers
In case you have decided to buy licenses, some of the top text analytics solutions are provided by the following providers - SAS Text Analytics; IBM Text Analytics; Lexalytics; Clarabridge and Attensity


Glossary
1.      IP - Internet Protocol
2.      IVR - Interactive Voice Response Unit
3.      NPS - Net Promoter Score
4.      ASAT - Agent Satisfaction Score
5.      FCR - First Call Resolution

Sources: Wikipedia.com, AllAnalytics.com, Gartner.com




Comments