- Implementing a full Transformer model with all the components
- Implementing a spam classifier using a pretrained BERT model from TFHub
- Implementing a question-answering model using Hugging Face’s Transformer library
13.1 Transformers in more detail
13.1.1 Revisiting the basic components of the Transformer
13.1.2 Embeddings in the Transformer
13.1.3 Residuals and normalization
13.2 Using pretrained BERT for spam classification
13.2.1 Understanding BERT
13.2.2 Classifying spam with BERT in TensorFlow
13.3 Question answering with Hugging Face’s Transformers
13.3.1 Understanding the data
13.3.3 Defining the DistilBERT model
13.3.4 Training the model
13.3.5 Ask BERT a question