This chapter covers
- Semantic search using Large Language Model (LLM) embeddings
- Representing the meaning of text with dense vectors
- An introduction to Transformers, and their impact on text representation and retrieval
- Building a fast and accurate autocomplete using Transformer models
- Using approximate nearest neighbor (ANN) search to speed up dense vector retrieval
In this chapter, we’ll start our journey into the emerging future of search, where the hyper-contextual vectors generated by Large Language Models (LLMs) are driving significant improvements to interpretation of queries, documents, and search results. Further, generative LLMs (like ChatGPT by OpenAI and many other commercial and open-source alternatives) are also able to use these vectors to generate new content, including query expansion, search training data, and summarization of search results, as we’ll explore further in the coming chapters.