11 Part 1: Sequence to sequence learning
This chapter covers,
- Recognize the nature and process sequence-to-sequence data, where one sequence of arbitrary length (e.g. a sentence) is mapped to another sequence of arbitrary length using machine learning
- Write a sequence-to-sequence machine translation model that consumes raw text and internally convert it to numbers
- Write a custom loop to train the model and evaluate the model using BLEU (BiLingual Evaluation Understudy) metric. BLEU is a metric used to measure the quality of generated text sequences (e.g. translations)
- Repurposing the trained model to build a recursive inference model that can generate translations for unseen text
In the previous chapter, we discussed solving an NLP task known as language modelling with deep recurrent neural networks. In this chapter, we are going to further our discussion and learn how we can use recurrent neural networks to solve more complex tasks. Namely, we will learn about a variety of tasks, where an arbitrary length input sequence is mapped to another arbitrary length sequence. Machine translation would be a very appropriate example of this, where you convert a sequence of words in one language to a sequence of words in another.