5 Improving weak understanding
This chapter covers
- Identifying the types of errors a classifier can make
- Establishing a baseline of current classifier performance
- Using data science methodologies to identify and prioritize improvements
- Infusing your traditional AI with generated content to enhance understanding
In this chapter, we will go on a journey of iterative improvement cycles to improve the understanding (or classifier performance) of a conversational solution. Although data science techniques are used, you do not need to be a data scientist to extract meaningful insights about your data using the methodologies presented in this chapter.
5.1 Building your improvement plan
If you built a blind test set using a sample from your production logs, you should have a reliable "representative distribution" test set. This means that the topics that are most frequently asked by your users are represented with corresponding volume in your testing data. This will be a key factor in prioritizing any issues that are surfaced by your test results.
If you are working with the results of a k-fold test (refer to Chapter 4), you won't know for certain which topics are the most important, so the most egregious accuracy scores are a logical starting point. In either case, it's now time to dig into those test results. An improvement plan starts with identifying the biggest problem spots in the bot’s training.