Chapter 3. Sequential labeling and language modeling

Chapter 10 from Real-World Natural Language Processing by Masato Hagiwara

This chapter covers

  • Solving part-of-speech (POS) tagging and named entity recognition (NER) using sequential labeling
  • Making RNNs more powerful—multi-layer and bidirectional recurrent neural networks (RNNs)
  • Capturing statistical properties of language using language models
  • Using language models to evaluate and generate natural language text

In this chapter, we are going to discuss sequential labeling—an important NLP framework where systems tag individual words with corresponding labels. Many NLP applications, such as part-of-speech tagging and named entity recognition, can be framed as sequential labeling tasks. In the second half of the chapter, I’ll introduce the concept of language models, one of the most fundamental yet exciting topics in NLP. I’ll talk about why they are important, how to use them to evaluate and even generate natural language text.

5.1 Introduction to sequential labeling

5.2 Building a part-of-speech tagger

5.3 Multi-layer and bidirectional RNNs

5.4 Named entity recognition

5.5 Modeling a language

5.6 Text generation using RNNs

5.7 Summary