2 Now we are speaking
This chapter covers
- Introducing the first computational approaches to language.
- Breaking text into machine-readable units.
- Mapping words into spaces where meaning takes geometric form.
- Exposing why language remains a demanding frontier for AI.
Humans rely on language to coordinate actions, share knowledge, and explain ideas to one another. Through speech and writing, information can travel far beyond the moment in which it was created, allowing knowledge to accumulate across generations and become part of a shared intellectual inheritance For machines, however, language has always posed a difficult challenge, because words rarely carry meaning on their own and instead depend on context, intent, and the surrounding conversation. Teaching a computer to process text therefore means confronting the full complexity of how people express ideas, from simple descriptions to abstract arguments that unfold across entire paragraphs.
This chapter follows the long arc from early computational attempts to model language to the representations that make modern systems possible. It explores how rules gave way to statistics, how words were broken into machine-readable units, and how those units were embedded into numerical representations that capture patterns of meaning. Along the way, it examines why language poses challenges that go beyond syntax or vocabulary: ambiguity, long-range dependencies, pragmatic inference, and the sheer diversity of human expression.