Abstract
The design, evaluation and adaptation of conversational information systems are typically guided by ratings from third-party, i.e. non-user, annotators. Interfaces used in gathering such ratings are designed in an ad-hoc fashion as it has not yet been investigated which design yields high-quality ratings. This work describes how to design user interfaces for gathering high-quality ratings with third-party annotators. In a user study, we compare a base interface that consolidates best practices from literature, an interface with clear definitions and an interface in which tasks are separated visually. We find that these interfaces yield annotations of high quality and separation of tasks. We find no significant improvements in quality between UIs. This work can serve as a starting point for researchers and practitioners interested in collecting high-quality dialogue user satisfaction ratings using third-party annotators.
Original language | English |
---|---|
Title of host publication | CHIIR 2020 |
Subtitle of host publication | Proceedings of the 2020 Conference on Human Information Interaction and Retrieval |
Publisher | ACM |
Pages | 363-367 |
Number of pages | 5 |
ISBN (Print) | 9781450368926 |
DOIs | |
Publication status | Published - Apr 2020 |
Keywords
- Human-centered Computing
- Natural language interfaces
- user interface design
- information systems
- search interfaces
VU Research Profile
- Connected World