TY - GEN
T1 - An uncertainty-aware query selection model for evaluation of IR systems
AU - Hosseini, Mehdi
AU - Cox, Ingemar J.
AU - Milić-Frayling, Nataša
AU - Shokouhi, Milad
AU - Yilmaz, Emine
PY - 2012
Y1 - 2012
N2 - We propose a mathematical framework for query selection as a mechanism for reducing the cost of constructing information retrieval test collections. In particular, our mathematical formulation explicitly models the uncertainty in the retrieval effectiveness metrics that is introduced by the absence of relevance judgments. Since the optimization problem is computationally intractable, we devise an adaptive query selection algorithm, referred to as Adaptive, that provides an approximate solution. Adaptive selects queries iteratively and assumes that no relevance judgments are available for the query under consideration. Once a query is selected, the associated relevance assessments are acquired and then used to aid the selection of subsequent queries. We demonstrate the effectiveness of the algorithm on two TREC test collections as well as a test collection of an online search engine with 1000 queries. Our experimental results show that the queries chosen by Adaptive produce reliable performance ranking of systems. The ranking is better correlated with the actual systems ranking than the rankings produced by queries that were selected using the considered baseline methods.
AB - We propose a mathematical framework for query selection as a mechanism for reducing the cost of constructing information retrieval test collections. In particular, our mathematical formulation explicitly models the uncertainty in the retrieval effectiveness metrics that is introduced by the absence of relevance judgments. Since the optimization problem is computationally intractable, we devise an adaptive query selection algorithm, referred to as Adaptive, that provides an approximate solution. Adaptive selects queries iteratively and assumes that no relevance judgments are available for the query under consideration. Once a query is selected, the associated relevance assessments are acquired and then used to aid the selection of subsequent queries. We demonstrate the effectiveness of the algorithm on two TREC test collections as well as a test collection of an online search engine with 1000 queries. Our experimental results show that the queries chosen by Adaptive produce reliable performance ranking of systems. The ranking is better correlated with the actual systems ranking than the rankings produced by queries that were selected using the considered baseline methods.
KW - information retrieval
KW - query selection
KW - test collection
UR - http://www.scopus.com/inward/record.url?scp=84866598080&partnerID=8YFLogxK
U2 - 10.1145/2348283.2348403
DO - 10.1145/2348283.2348403
M3 - Conference contribution
AN - SCOPUS:84866598080
SN - 9781450316583
T3 - SIGIR'12 - Proceedings of the International ACM SIGIR Conference on Research and Development in Information Retrieval
SP - 901
EP - 910
BT - SIGIR'12 - Proceedings of the International ACM SIGIR Conference on Research and Development in Information Retrieval
T2 - 35th Annual ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2012
Y2 - 12 August 2012 through 16 August 2012
ER -