Show simple item record

AuthorKutlu M.
AuthorElsayed T.
AuthorHasanain M.
AuthorLease M.
Available date2020-02-24T08:57:14Z
Publication Date2018
Publication NameInternational Conference on Information and Knowledge Management, Proceedings
ResourceScopus
URIhttp://dx.doi.org/10.1145/3269206.3271751
URIhttp://hdl.handle.net/10576/13016
AbstractBecause it is expensive to construct test collections for Cranfield-based evaluation of information retrieval systems, a variety of lower-cost methods have been proposed. The reliability of these methods is often validated by measuring rank correlation (e.g., Kendall's t) between known system rankings on the full test collection vs. observed system rankings on the lower-cost one. However, existing rank correlation measures do not consider the statistical significance of score differences between systems in the observed rankings. To address this, we propose two statistical-significance-aware rank correlation measures, one of which is a head-weighted version of the other. We first show empirical differences between our proposed measures and existing ones. We then compare the measures while benchmarking four system evaluation methods: pooling, crowdsourcing, evaluation with incomplete judgments, and automatic system ranking. We show that use of our measures can lead to different experimental conclusions regarding reliability of alternative low-cost evaluation methods.
SponsorThis work was made possible by NPRP grant# NPRP 7-1313-1-245 from the Qatar National Research Fund (a member of Qatar Foundation). The statements made herein are solely the responsibility of the authors.
Languageen
PublisherAssociation for Computing Machinery
SubjectEvaluation
IR System Ranking
Rank Correlation
TitleWhen rank order isn't enough: New statistical-significance-aware correlation measures
TypeConference Paper
Pagination397 - 406


Files in this item

FilesSizeFormatView

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record