TY - GEN
T1 - Towards Credible Human Evaluation of Open-Domain Dialog Systems Using Interactive Setup
AU - Liu, Sijia
AU - Lange, Patrick
AU - Hedayatnia, Behnam
AU - Papangelis, Alexandros
AU - Jin, Di
AU - Wirth, Andrew
AU - Liu, Yang
AU - Hakkani-Tur, Dilek
N1 - We would like to thank Karthik Gopalakrishnan for his initial passion and ideas with this work, and Seokhwan Kim and Shikib Mehri for their inspiring feedback. We would also like to especially thank Prof. Joel Greenhouse from Department of Statistics & Data Science, Carnegie Mellon University, for his feedback and suggestions over the statistical methodology covered in this project.
PY - 2023/6/27
Y1 - 2023/6/27
N2 - Evaluating open-domain conversation models has been an open challenge due to the open-ended nature of conversations. In addition to static evaluations, recent work has started to explore a variety of per-turn and per-dialog interactive evaluation mechanisms and provide advice on the best setup. In this work, we adopt the interactive evaluation framework and further apply to multiple models with a focus on per-turn evaluation techniques. Apart from the widely used setting where participants select the best response among different candidates at each turn, one more novel per-turn evaluation setting is adopted, where participants can select all appropriate responses with different fallback strategies to continue the conversation when no response is selected. We evaluate these settings based on sensitivity and consistency using four GPT2-based models that differ in model sizes or fine-tuning data. To better generalize to any model groups with no prior assumptions on their rankings and control evaluation costs for all setups, we also propose a methodology to estimate the required sample size given a minimum performance gap of interest before running most experiments. Our comprehensive human evaluation results shed light on how to conduct credible human evaluations of open domain dialog systems using the interactive setup, and suggest additional future directions.
AB - Evaluating open-domain conversation models has been an open challenge due to the open-ended nature of conversations. In addition to static evaluations, recent work has started to explore a variety of per-turn and per-dialog interactive evaluation mechanisms and provide advice on the best setup. In this work, we adopt the interactive evaluation framework and further apply to multiple models with a focus on per-turn evaluation techniques. Apart from the widely used setting where participants select the best response among different candidates at each turn, one more novel per-turn evaluation setting is adopted, where participants can select all appropriate responses with different fallback strategies to continue the conversation when no response is selected. We evaluate these settings based on sensitivity and consistency using four GPT2-based models that differ in model sizes or fine-tuning data. To better generalize to any model groups with no prior assumptions on their rankings and control evaluation costs for all setups, we also propose a methodology to estimate the required sample size given a minimum performance gap of interest before running most experiments. Our comprehensive human evaluation results shed light on how to conduct credible human evaluations of open domain dialog systems using the interactive setup, and suggest additional future directions.
UR - https://www.scopus.com/pages/publications/85167998404
UR - https://www.scopus.com/inward/citedby.url?scp=85167998404&partnerID=8YFLogxK
U2 - 10.1609/aaai.v37i11.26557
DO - 10.1609/aaai.v37i11.26557
M3 - Conference contribution
AN - SCOPUS:85167998404
T3 - Proceedings of the 37th AAAI Conference on Artificial Intelligence, AAAI 2023
SP - 13264
EP - 13272
BT - AAAI-23 Technical Tracks 11
A2 - Williams, Brian
A2 - Chen, Yiling
A2 - Neville, Jennifer
PB - American Association for Artificial Intelligence (AAAI) Press
T2 - 37th AAAI Conference on Artificial Intelligence, AAAI 2023
Y2 - 7 February 2023 through 14 February 2023
ER -