RMIT University
Browse

On the effect of relevance scales in crowdsourcing relevance assessments for Information Retrieval evaluation

journal contribution
posted on 2024-11-02, 18:35 authored by Kevin Roitero, Eddy Maddalena, Stefano Mizzaro, Falk ScholerFalk Scholer
Relevance is a key concept in information retrieval and widely used for the evaluation of search systems using test collections. We present a comprehensive study of the effect of the choice of relevance scales on the evaluation of information retrieval systems. Our work analyzes and compares four crowdsourced scales (2-levels, 4-levels, and 100-levels ordinal scales, and a magnitude estimation scale) and two expert-labeled datasets (on 2- and 4-levels ordinal scales). We compare the scales considering internal and external agreement, the effect on IR evaluation both in terms of system effectiveness and topic ease, and we discuss the effect of such scales and datasets on the perception of relevance levels by assessors. Our analyses show that: crowdsourced judgment distributions are consistent across scales, both overall and at the per-topic level; on all scales crowdsourced judgments agree with the expert judgments, and overall the crowd assessors are able to express reliable relevance judgments; all scales lead to a similar level of external agreement with the ground truth, while the internal agreement among crowd workers is higher for fine-grained scales; more fine-grained scales consistently lead to higher correlation values for both system ranking and topic ease; finally, we found that the considered scales lead to different perceived distances between relevance levels.

Funding

New approaches to interactive sessional search for complex tasks

Australian Research Council

Find out more...

History

Related Materials

  1. 1.
    DOI - Is published in 10.1016/j.ipm.2021.102688
  2. 2.
    ISSN - Is published in 03064573

Journal

Information Processing and Management

Volume

58

Number

102688

Issue

6

Start page

1

End page

23

Total pages

23

Publisher

Elsevier

Place published

United Kingdom

Language

English

Copyright

© 2021 Elsevier Ltd. All rights reserved.

Former Identifier

2006110833

Esploro creation date

2021-11-21

Usage metrics

    Scholarly Works

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC