« Back to publications

Extended Overview of the CLEF-2023 LongEval Lab on Longitudinal Evaluation of Model Performance

Rabab Alkhalifa, Iman Bilal, Hsuvas Borkakoty, Jose Camacho-Collados, Romain Deveaud, Alaa El-Ebshihy, Luis Espinosa-Anke, Gabriela Gonzalez Saez, Petra GalušÄáková, Lorraine Goeuriot, Elena Kochkina, Maria Liakata, Daniel Loureiro, Philippe Mulhem, Florina Piroi, Martin Popel, Christophe Servan, Harish Tayyar Madabushi, Arkaitz Zubiaga

LongEval. 2023.

Download PDF file
We describe the first edition of the LongEval CLEF 2023 shared task. This lab evaluates the temporal persistence of Information Retrieval (IR) systems and Text Classifiers. Task 1 requires IR systems to run on corpora acquired at several timestamps, and evaluates the drop in system quality (NDCG) along these timestamps. Task 2 tackles binary sentiment classification at different points in time, and evaluates the performance drop for different temporal gaps. Overall, 37 teams registered for Task 1 and 25 for Task 2. Ultimately, 14 and 4 teams participated in Task 1 and Task 2, respectively.
@inproceedings{alkhalifa2023extended,
  title={Extended Overview of the CLEF-2023 LongEval Lab on Longitudinal Evaluation of Model Performance},
  author={Alkhalifa, Rabab and Bilal, Iman and Borkakoty, Hsuvas and Camacho-Collados, Jose and Deveaud, Romain and El-Ebshihy, Alaa and Espinosa-Anke, Luis and Gonzalez-Saez, Gabriela and Galu{\v{s}}{\v{c}}{\'a}kov{\'a}, Petra and Goeuriot, Lorraine and others},
  booktitle={CEUR Workshop Proceedings},
  volume={3497},
  pages={2181--2203},
  year={2023},
  organization={CEUR-WS}
}