5 Deep Transfer Learning for NLP with Transformers
This chapter covers:
- Understanding the basics of the transformer neural network architecture
- Using the generative pretrained transformer (GPT) to generate text
- Using pre-trained bidirectional encoder representations from transformers (BERT) architecture to perform some interesting tasks.
- Using the BERT architecture for cross-lingual transfer learning
In this chapter, we will cover some representative deep transfer learning modeling architectures for NLP that rely on a recently popularized neural architecture – the transformer[1] – for key functions. This is arguably the most important architecture for natural language processing (NLP) today. Specifically, we will be looking at modeling frameworks such as the generative pretrained transformer (GPT)[2], bidirectional encoder representations from transformers (BERT)[3] and multilingual BERT (mBERT)[4]. These methods employ neural networks with even more parameters than the deep convolutional and recurrent neural network models that we looked at in the previous chapter. Despite the larger size, they have exploded in popularity because they scale comparatively more effectively on parallel computing architecture. This enables even larger and more sophisticated models to be developed in practice.