chapter six
6 Beyond Natural Language Processing
This chapter covers
- Transformer layers can work on data that isn’t text
- LLMs need help writing working software
- LLMs can understand mathematical notation with some tweaks
- Transformers work with images by replacing the input and output steps
While modeling natural language was the transformers’ primary purpose, machine learning researchers quickly discovered they could predict anything involving data sequences. Transformers view a sentence as a sequence of tokens and either produce a related sequence of tokens, such as a translation from one language to another or predict the following tokens in a sequence, such as when answering questions or acting like a chatbot. While sequence modeling and prediction are potent tools for interpreting and generating natural language, natural language is the only domain where LLMs can be helpful.