RMIT University
Browse

Gauging the Quality of Relevance Assessments using Inter-Rater Agreement

conference contribution
posted on 2024-10-31, 20:50 authored by Tadele Tedla Damessie, Jessie Nghiem, Falk ScholerFalk Scholer, Shane CulpepperShane Culpepper
In recent years, gathering relevance judgments through non-topic originators has become an increasingly important problem in Information Retrieval. Relevance judgments can be used to measure the effectiveness of a system, and are often needed to build supervised learning models in learning-to-rank retrieval systems. The two most popular approaches to gathering bronze level judgments - where the judge is not the originator of the information need for which relevance is being assessed, and is not a topic expert - is through a controlled user study, or through crowdsourcing. However, judging comes at a cost (in time, and usually money) and the quality of the judgments can vary widely. In this work, we directly compare the reliability of judgments using three different types of bronze assessor groups. Our first group is a controlled Lab group; the second and third are two different crowdsourcing groups, CF-Document where assessors were free to judge any number of documents for a topic, and CF-Topic where judges were required to judge all of the documents from a single topic, in a manner similar to the Lab group. Our study shows that Lab assessors exhibit a higher level of agreement with a set of ground truth judgments than CF-Topic and CF-Document assessors. Inter-rater agreement rates show analogous trends. These finding suggests that in the absence of ground truth data, agreement between assessors can be used to reliably gauge the quality of relevance judgments gathered from secondary assessors, and that controlled user studies are more likely to produce reliable judgments despite being more costly.

Funding

Trajectory data processing: Spatial computing meets information retrieval

Australian Research Council

Find out more...

History

Related Materials

  1. 1.
    DOI - Is published in 10.1145/3077136.3080729
  2. 2.
    ISBN - Is published in 9781450350228 (urn:isbn:9781450350228)

Start page

1089

End page

1092

Total pages

4

Outlet

Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval 2017

Name of conference

SIGIR 2017

Publisher

ACM

Place published

New York, New York

Start date

2017-08-07

End date

2017-08-11

Language

English

Former Identifier

2006076331

Esploro creation date

2020-06-22

Fedora creation date

2017-08-15

Usage metrics

    Scholarly Works

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC