chapter seven
7 The Pivot to Reasoning
This chapter covers
- VLAE: lossy latents with a PixelCNN decoder and an autoregressive prior
- Relation Networks: pairwise relational reasoning on CLEVR and bAbI
- Message Passing Neural Networks: graph learning and chemical accuracy on QM9
- Relational Memory Core: attention among memory slots for sequential reasoning
- Limits and Living doubts
Scaling pretraining gave us more coherent text generation, but it did not provide dependable multi-step reasoning or planning. OpenAI’s trajectory after GPT-4 (2023) reflects this reality. Rather than scaling to GPT-5 with trillions of parameters, OpenAI invested in techniques like “test-time compute” that involve models “thinking” more at inference and fine-tuning with human feedback, essentially seeking algorithmic and architectural improvements beyond brute force.