RMIT University
Browse

Extending test collection pools without manual runs

conference contribution
posted on 2024-10-31, 17:46 authored by Gaya Jayasinghe, William Webber, Mark SandersonMark Sanderson, Shane CulpepperShane Culpepper
Information retrieval test collections traditionally use a combination of automatic and manual runs to create a pool of documents to be judged. The quality of the final judgments produced for a collection is a product of the variety across each of the runs submitted and the pool depth. In this work, we explore fully automated approaches to generating a pool. By combining a simple voting approach with machine learning from documents retrieved by automatic runs, we are able to identify a large portion of relevant documents that would normally only be found through manual runs. Our initial results are promising and can be extended in future studies to help test collection curators ensure proper judgment coverage is maintained across complete document collections.

History

Related Materials

  1. 1.
    DOI - Is published in 10.1145/2600428.2609473
  2. 2.
    ISBN - Is published in 9781450322577 (urn:isbn:9781450322577)

Start page

915

End page

918

Total pages

4

Outlet

Proceedings of 37th International ACM SIGIR Conference on Research and Development in Information Retrieval

Editors

Shlomo Geva, Andrew Trotman, Peter Bruza, Charles L.A. Clarke, Kal Järvelin

Name of conference

SIGIR 2014

Publisher

Association for Computing Machinery

Place published

New York, United States

Start date

2014-07-06

End date

2014-07-11

Language

English

Copyright

Copyright is held by the owner/author(s). Publication rights licensed to ACM.

Former Identifier

2006047724

Esploro creation date

2020-06-22

Fedora creation date

2014-11-10

Usage metrics

    Scholarly Works

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC