This chapter covers
- Building a question-answering application
- Curating a question-answering dataset for training
- Fine-tuning a transformer model
- Blending retrieval strategies by integrating deep-learning-based NLP with Solr
- The future of AI-powered search and emerging search paradigms
With the basics of semantic search with transformers well understood from Chapter 13, we’re now ready to attempt one of the hardest problems in search: Question Answering. This problem, while a lofty goal, was chosen for several key reasons: it will (a) help you better understand the transformers tooling and ecosystem, (b) teach fine-tuning of large language models to a specific task, and (c) merge the Solr search engine and advanced natural language techniques together to produce a complete solution.
With our question-answering application in hand, we will then touch on how else search is evolving and what to expect in the coming years.
In this section we will introduce the question-answering problem space and provide an overview of the retriever-reader pattern for implementing question answering.
Traditional search returns lists of documents or pages in response to a query, but often people may just be looking for a quick answer to their question versus wanting to spend time reviewing the underlying documents.