CENTRE@CLEF 2018

CLEF   NTCIR   TREC
REproducibility

Aims and Scope

The goal of CENTRE@CLEF 2018 is to run a joint CLEF/NTCIR/TREC task on challenging participants to reproduce best results of the most interesting systems submitted in previous editions of CLEF/NTCIR/TREC and to contribute back to the community the additional components and resources developed to reproduce the results (lab flyer available here).

The CENTRE@CLEF 2018 lab will offer two pilot tasks:

  • Task 1 - Replicability: the task will focus on the replicability of selected methods on the same experimental collections;
  • Task 2 - Reproducibility: the task will focus on the reproducibility of selected methods on the different experimental collections.

Since CENTRE is a joint CLEF/NTCIR/TREC activity, participants in Tasks 1 and 2 will be challenged to reproduce methods and systems developed in all the evaluation campaigns, i.e. CLEF will challenge against NTCIR and TREC results in addition to CLEF results.

To participate in the lab, the groups need to register (by April, 27) at the following link:

Sign Up

Important Dates

Registration closes: April 27, 2018

Runs due from participants: May 11, 2018

Submission of participant papers: May 31, 2018

Notification of acceptance: June 15, 2018

Camera ready due: June 29, 2018

CLEF 2018 conference: September 10-14, 2018

Tasks Description

Among the systems submitted to the CLEF/NTCIR/TREC ad-hoc tasks we selected the following papers:

The following table reports, for each paper, the name of the run to be replicated and/or reproduced and the datasets and topics to be used for the replicability and reproducibility tasks.

Paper Run Name Replicability Task Reproducibility Task
[Guyot et al, 2005] AUTOEN Multi-8 Two Years On with topics of CLEF 2005, Ad Hoc Track, Multilingual Task Multi-8 Two Years On with topics of CLEF 2003, Ad Hoc Track, Multilingual Task
Multi-8 Two Years On with topics of CLEF 2004, Ad Hoc Track, Multilingual Task
[Nguyen et al, 2008] 10.2415/AH-TEL-BILI-X2EN-CLEF2008.TWENTE.FCW TEL English (BL) with topics of CLEF 2008, Ad Hoc Track, TEL Bilingual English Task TEL French (BNF) and TEL German (ONB) with topics of CLEF 2008, Ad Hoc Track, TEL Bilingual French and German Tasks
TEL English (BL), TEL French (BNF) and TEL German (ONB) with topics of CLEF 2009, Ad Hoc Track, TEL Bilingual English, French, and German Tasks
[Sorg et al, 2008] 10.2415/AH-TEL-BILI-X2DE-CLEF2008.KARLSRUHE.AIFB_ONB_EN TEL German (ONB) with topics of CLEF 2008, Ad Hoc Track, TEL Bilingual German Task TEL English (BL) and TEL French (BNF) with topics of CLEF 2008, Ad Hoc Track, TEL Bilingual English and German Tasks
TEL English (BL), TEL French (BNF) and TEL German (ONB) with topics of CLEF 2009, Ad Hoc Track, TEL Bilingual English, French, and German Tasks
[Yang et al, 2013] UDInfolabWEB2 ClueWeb12 Category A with topics of TREC 2013, Web Track, Ad Hoc Task ClueWeb09 Category A and B with topics of TREC 2012, Web Track, Ad Hoc Task
ClueWeb12 Category B with topics of TREC 2013, Web Track, Ad Hoc Task
ClueWeb12 Category A and B with topics of TREC 2014, Web Track, Ad Hoc Task
[McCreadie et al, 2014] uogTrDwl ClueWeb12 Category A with topics of TREC 2014, Web Track, Ad Hoc Task ClueWeb09 Category A and B with topics of TREC 2012, Web Track, Ad Hoc Task
ClueWeb12 Category A and B with topics of TREC 2013, Web Track, Ad Hoc Task
ClueWeb12 Category B with topics of TREC 2014, Web Track, Ad Hoc Task
[Gallagher et al, 2017] RMIT-E-NU-Own-1
RMIT-E-NU-Own-3
ClueWeb12 Category A with topics of NTCIR-13, We Want Web Track ClueWeb12 Category B with topics of NTCIR-13, We Want Web Track

Corpora:

CLEF datasets can be downloaded here. To get the username and password necessary to download the data, the participants need to follow the instruction written on the end user agreement.

TREC datasets are distributed by CMU at the link: http://lemurproject.org/.

Topics and Qrels:

CLEF topics and qrels are available here. To get the username and password necessary to download the data, the participants need to follow the instruction written on the end user agreement.

Topics and qrels needed for the Replicability Task:

Topics and qrels needed for the Reproducibility Task:

Submission Guidelines

Groups should satisfy the following guidelines to participate in this lab:

Trec Format:

Runs should be submitted with the following format:


30 Q0 ZF08-175-870  0 4238 prise1
30 Q0 ZF08-306-044  1 4223 prise1
30 Q0 ZF09-477-757  2 4207 prise1
30 Q0 ZF08-312-422  3 4194 prise1
30 Q0 ZF08-013-262  4 4189 prise1
...
   						
where: It is important to include all the columns and have a white space delimiter between the columns.

Runs should be submitted at the following link:

TBA

Evaluation:

The quality of the submitted runs will be evaluated from two point of views:

Organizers

Nicola Ferro, University of Padua, Italy
ferrodei.unipd.it

Maria Maistro, University of Padua, Italy
maistrodei.unipd.it

Tetsuya Sakai, Waseda University, Japan
tetsuyasakaiacm.org

Ian Soboroff, National Institute of Standards and Technology (NIST), US
ian.soboroffnist.gov