TY - GEN
T1 - Evaluation of methods for relative comparison of retrieval systems based on clickthroughs
AU - He, Jing
AU - Zhai, Chengxiang
AU - Li, Xiaoming
PY - 2009
Y1 - 2009
N2 - The Cranfield evaluation method has some disadvantages, including its high cost in labor and inadequacy for evaluating interactive retrieval techniques. As a very promising alternative, automatic comparison of retrieval systems based on observed clicking behavior of users has recently been studied. Several methods have been proposed, but there has so far been no systematic way to assess which strategy is better, making it difficult to choose a good method for real applications. In this paper, we propose a general way to evaluate these relative comparison methods with two measures: utility to users(UtU) and effectiveness of differentiation(EoD). We evaluate two state of the art methods by systematically simulating different retrieval scenarios. Inspired by the weakness of these methods revealed through our evaluation, we further propose a novel method by considering the positions of clicked documents. Experiment results show that our new method performs better than the existing methods.
AB - The Cranfield evaluation method has some disadvantages, including its high cost in labor and inadequacy for evaluating interactive retrieval techniques. As a very promising alternative, automatic comparison of retrieval systems based on observed clicking behavior of users has recently been studied. Several methods have been proposed, but there has so far been no systematic way to assess which strategy is better, making it difficult to choose a good method for real applications. In this paper, we propose a general way to evaluate these relative comparison methods with two measures: utility to users(UtU) and effectiveness of differentiation(EoD). We evaluate two state of the art methods by systematically simulating different retrieval scenarios. Inspired by the weakness of these methods revealed through our evaluation, we further propose a novel method by considering the positions of clicked documents. Experiment results show that our new method performs better than the existing methods.
KW - Evaluation
KW - Implicit feedback
KW - Information retrieval
UR - http://www.scopus.com/inward/record.url?scp=74549144985&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=74549144985&partnerID=8YFLogxK
U2 - 10.1145/1645953.1646293
DO - 10.1145/1645953.1646293
M3 - Conference contribution
AN - SCOPUS:74549144985
SN - 9781605585123
T3 - International Conference on Information and Knowledge Management, Proceedings
SP - 2029
EP - 2032
BT - ACM 18th International Conference on Information and Knowledge Management, CIKM 2009
T2 - ACM 18th International Conference on Information and Knowledge Management, CIKM 2009
Y2 - 2 November 2009 through 6 November 2009
ER -