TY - GEN
T1 - CheckThat! at CLEF 2020
T2 - 42nd European Conference on IR Research, ECIR 2020
AU - Barrón-Cedeño, Alberto
AU - Elsayed, Tamer
AU - Nakov, Preslav
AU - Da San Martino, Giovanni
AU - Hasanain, Maram
AU - Suwaileh, Reem
AU - Haouari, Fatima
N1 - Publisher Copyright:
© Springer Nature Switzerland AG 2020.
PY - 2020
Y1 - 2020
N2 - We describe the third edition of the CheckThat! Lab, which is part of the 2020 Cross-Language Evaluation Forum (CLEF). CheckThat! proposes four complementary tasks and a related task from previous lab editions, offered in English, Arabic, and Spanish. Task 1 asks to predict which tweets in a Twitter stream are worth fact-checking. Task 2 asks to determine whether a claim posted in a tweet can be verified using a set of previously fact-checked claims. Task 3 asks to retrieve text snippets from a given set of Web pages that would be useful for verifying a target tweet’s claim. Task 4 asks to predict the veracity of a target tweet’s claim using a set of potentially-relevant Web pages. Finally, the lab offers a fifth task that asks to predict the check-worthiness of the claims made in English political debates and speeches. CheckThat! features a full evaluation framework. The evaluation is carried out using mean average precision or precision at rank k for ranking tasks, and F1 for classification tasks.
AB - We describe the third edition of the CheckThat! Lab, which is part of the 2020 Cross-Language Evaluation Forum (CLEF). CheckThat! proposes four complementary tasks and a related task from previous lab editions, offered in English, Arabic, and Spanish. Task 1 asks to predict which tweets in a Twitter stream are worth fact-checking. Task 2 asks to determine whether a claim posted in a tweet can be verified using a set of previously fact-checked claims. Task 3 asks to retrieve text snippets from a given set of Web pages that would be useful for verifying a target tweet’s claim. Task 4 asks to predict the veracity of a target tweet’s claim using a set of potentially-relevant Web pages. Finally, the lab offers a fifth task that asks to predict the check-worthiness of the claims made in English political debates and speeches. CheckThat! features a full evaluation framework. The evaluation is carried out using mean average precision or precision at rank k for ranking tasks, and F1 for classification tasks.
UR - http://www.scopus.com/inward/record.url?scp=85084188926&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-45442-5_65
DO - 10.1007/978-3-030-45442-5_65
M3 - Conference contribution
AN - SCOPUS:85084188926
SN - 9783030454418
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 499
EP - 507
BT - Advances in Information Retrieval - 42nd European Conference on IR Research, ECIR 2020, Proceedings
A2 - Jose, Joemon M.
A2 - Yilmaz, Emine
A2 - Magalhães, João
A2 - Martins, Flávio
A2 - Castells, Pablo
A2 - Ferro, Nicola
A2 - Silva, Mário J.
PB - Springer
Y2 - 14 April 2020 through 17 April 2020
ER -