TY - JOUR
T1 - Overview of the CLEF-2025 CheckThat! Lab Task 3 on Fact-Checking Numerical Claims
AU - Venktesh, V.
AU - Setty, Vinay
AU - Anand, Avishek
AU - Bendou, Boushra
AU - Hasanain, Maram
AU - Bouamor, Houda
AU - Iturra-Bocaz, Gabriel
AU - Galuščáková, Petra
AU - Alam, Firoj
N1 - Publisher Copyright:
© 2025 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
PY - 2025
Y1 - 2025
N2 - We present an overview of the CheckThat! Lab 2025 Task 3, part of CLEF 2025. The task focuses on verifying claims with numerical quantities and temporal expressions. Numerical claims are defined as those requiring validation of explicit or implicit quantitative or temporal details. It is conducted in three languages: Arabic, Spanish, and English. A total of 258 valid runs were submitted by 13 unique teams across languages, with 4 participants in Spanish and Arabic. 10 teams participated in fact-checking English numerical claims. Among these teams, the use of transformer pre-trained language models (PLMs) was the most frequent. A few teams also employed Large Language Models (LLMs). We provide a description of the dataset, the task setup, including evaluation settings, and a brief overview of the participating systems. As is customary in the CheckThat! Lab, we release all the datasets as well as the evaluation scripts to the research community. This will enable further research on identifying challenges with fact-checking numerical claims that can assist various stakeholders, such as fact-checkers, financial research analysts, and policymakers.
AB - We present an overview of the CheckThat! Lab 2025 Task 3, part of CLEF 2025. The task focuses on verifying claims with numerical quantities and temporal expressions. Numerical claims are defined as those requiring validation of explicit or implicit quantitative or temporal details. It is conducted in three languages: Arabic, Spanish, and English. A total of 258 valid runs were submitted by 13 unique teams across languages, with 4 participants in Spanish and Arabic. 10 teams participated in fact-checking English numerical claims. Among these teams, the use of transformer pre-trained language models (PLMs) was the most frequent. A few teams also employed Large Language Models (LLMs). We provide a description of the dataset, the task setup, including evaluation settings, and a brief overview of the participating systems. As is customary in the CheckThat! Lab, we release all the datasets as well as the evaluation scripts to the research community. This will enable further research on identifying challenges with fact-checking numerical claims that can assist various stakeholders, such as fact-checkers, financial research analysts, and policymakers.
UR - https://www.scopus.com/pages/publications/105019048155
M3 - Conference article
AN - SCOPUS:105019048155
SN - 1613-0073
VL - 4038
SP - 709
EP - 721
JO - CEUR Workshop Proceedings
JF - CEUR Workshop Proceedings
T2 - 26th Working Notes of the Conference and Labs of the Evaluation Forum, CLEF 2025
Y2 - 9 September 2025 through 12 September 2025
ER -