12 Overcoming ranking bias through active learning
This chapter covers
- Harnessing live user interactions to gather feedback on a deployed LTR model
- A/B testing search relevance solutions with live users
- Using active learning to explore potentially relevant results beyond the top results
- Balancing exploiting user interactions while exploring what else might be relevant
So far, our learning to rank (LTR) work has taken place in the lab. In previous chapters, we built models using automatically constructed training data from user clicks. In this chapter, we’ll take our model into the real world for a test drive with (simulated) live users!
Recall that we compared an automated LTR system to a self-driving car. Internally, the car has an engine: the end-to-end model retraining on historical judgments as discussed in chapter 10. In chapter 11, we compared our model’s training data to self-driving car directions: what should we optimize to automatically learn judgments based on previous interactions with search results? We built training data and overcame key biases inherent in click data.