Topic modeling; User Study; Mixed-Initiative Interaction;
Topic modeling is a common tool for understanding large bodies of text, but is typically provided as a “take it or leave it” proposition. Incorporating human knowledge in unsupervised learning is a promising approach to create high-quality topic models. Existing interactive systems and modeling algorithms support a wide range of refinement operations to express feedback. However, these systems’ interactions are primarily driven by algorithmic convenience, ignoring users who may lack expertise in topic modeling. To better understand how non-expert users understand, assess, and refine topics, we conducted two user studies—an in-person interview study and an online crowdsourced study. These studies demonstrate a disconnect between what non-expert users want and the complex, low-level operations that current interactive systems support. In particular, our findings include: (1) analysis of how non-expert users perceive topic models; (2) characterization of primary refinement operations expected by non-expert users and ordered by relative preference; (3) further evidence of the benefits of supporting users in directly refining a topic model; (4) design implications for future human-in-the-loop topic modeling interfaces.
Original Publication Citation
Lee, Tak Yeon, et al. "The Human Touch: How Non-expert Users Perceive, Interpret, and Fix Topic Models." International Journal of Human-Computer Studies (2017).
BYU ScholarsArchive Citation
Lee, Tak Yeon; Smith, Alison; Seppi, Kevin; Elmqvist, Niklas; Boyd-Graber, Jordan; and Findlater, Leah, "The Human Touch: How Non-expert Users Perceive, Interpret, and Fix Topic Models" (2017). All Faculty Publications. 1847.
Physical and Mathematical Sciences
© 2016 Elsevier Ltd see www.elsevier.com/copyright. Final publisher's version can be found at: http://dx.doi.org/10.1016/j.ijhcs.2017.03.007
Copyright Use Information
Available for download on Sunday, September 01, 2019