chapter seven

7 Selecting the Right People to Annotate your Data

 

This chapter covers

  • Understanding the characteristics of the three main types of annotation workforces: in-house, contracted, and pay-per-task.
  • Applying the three key principles for motivating an annotator - salary, security & transparency - and how to apply the principles to different workforces.
  • Evaluating workforces with non-monetary compensation systems, including application end-users, volunteers, people playing games, and computer-generated data/annotations.
  • Evaluating your annotation volume requirements, in order to understand the scale and frequency of the annotations that you will need.
  • Deciding on the qualifications necessary for an annotation task, in order to understand the amount of training and/or expertise needed by annotators.

In the last few chapters of the book, you learned how to select the right data for human review. Now, the following chapters will cover how to optimize that human interaction. Machine Learning models often require thousands (and sometimes millions) of instances of human feedback in order to get the training data necessary to be accurate.

7.1       Introduction to data labeling

7.1.1   Three principles for good data annotation

7.1.2   Annotating data and reviewing Machine Learning predictions

7.1.3   Annotations from Machine Learning assisted humans

7.2       In-house experts 

7.2.1   Salary for in-house workers

7.2.2   Security for in-house workers

7.2.3   Ownership for in-house workers

7.2.4   Tip: Always run in-house annotation sessions

7.3       Outsourced workers

7.3.1   Salary for outsourced workers

7.3.2   Security for outsourced workers

7.3.3   Ownership for outsourced workers

7.3.4   Tip: Talk to your outsourced workers

7.4       Crowdsourced workers

7.4.1   Salary for crowdsourced workers

7.4.2   Security for crowdsourced workers

7.4.3   Ownership for crowdsourced workers

7.4.4   Tip: Create a path to secure work and career advancement