chapter two
2 Data Privacy and Safety
This chapter covers
- Training large language models (LLMs) with open web data collection, autoregression and bidirectional token prediction, and fine-tuning.
- The potential emergent abilities, harms, and vulnerabilities that come from training LLMs.
- Improving “desirable” outputs from LLMs: post-processing detection algorithms, content filtering or conditional pre-training, reinforcement learning from human feedback (RLHF), and constitutional AI or reinforcement learning from AI feedback (RLAIF).
- Mitigating privacy risks with user inputs to chatbots.
- Understanding data protection laws in the U.S. and the European Union (EU).
For decades, the digital economy has run on the currency of data. The digital economy of collecting and trading information about who we are and what we do online is worth trillions of dollars, and as more of our daily activities have moved on to the internet, the mill has ever more grist to grind through. Large language models are inventions of the internet age, emulating human language by vacuuming up terabytes[10] of text data found online.