Abstract
The labeling of language resources is a time consuming task, whether aided by machine learning or not. Much of the prior work in this area has focused on accelerating human annotation in the context of machine learning, yielding a variety of active learning approaches. Most of these attempt to lead an annotator to label the items which are most likely to improve the quality of an automated, machine learning-based model. These active learning approaches seek to understand the effect of item selection on the machine learning model, but give significantly less emphasis to the effect of item selection on the human annotator. In this work, we consider a sentiment labeling task where existing, traditional active learning seems to have little or no value. We focus instead on the human annotator by ordering the items for better annotator efficiency.
Degree
MS
College and Department
Physical and Mathematical Sciences; Computer Science
Rights
http://lib.byu.edu/about/copyright/
BYU ScholarsArchive Citation
Okuda, Nozomu, "The Annotation Cost of Context Switching: How Topic Models and Active Learning [May Not] Work Together" (2017). Theses and Dissertations. 6906.
https://scholarsarchive.byu.edu/etd/6906
Date Submitted
2017-08-01
Document Type
Thesis
Handle
http://hdl.lib.byu.edu/1877/etd9454
Keywords
active learning, topic modeling, annotation, human cost
Language
english