TY - JOUR
T1 - Overview of the CLEF-2018 CheckThat! Lab on Automatic Identification and Verification of Political Claims. Task 1
T2 - 19th Working Notes of CLEF Conference and Labs of the Evaluation Forum, CLEF 2018
AU - Atanasova, Pepa
AU - Màrquez, Lluís
AU - Barrón-Cedeño, Alberto
AU - Elsayed, Tamer
AU - Suwaileh, Reem
AU - Zaghouani, Wajdi
AU - Kyuchukov, Spas
AU - Da San Martino, Giovanni
AU - Nakov, Presiav
PY - 2018
Y1 - 2018
N2 - We present an overview of the CLEF-2018 CheckThat! Lab on Automatic Identification and Verification of Political Claims, with focus on Task 1: Check-Worthiness. The task asks to predict which claims in a political debate should be prioritized for fact-checking. In particular, given a debate or a political speech, the goal was to produce a ranked list of its sentences based on their worthiness for fact checking. We offered the task in both English and Arabic, based on debates from the 2016 US Presidential Campaign, as well as on some speeches during and after the campaign. A total of 30 teams registered to participate in the Lab and seven teams actually submitted systems for Task 1. The most successful approaches used by the participants relied on recurrent and multi-layer neural networks, as well as on combinations of distributional representations, on matchings claims' vocabulary against lexicons, and on measures of syntactic dependency. The best systems achieved mean average precision of 0.18 and 0.15 on the English and on the Arabic test datasets, respectively. This leaves large room for further improvement, and thus we release all datasets and the scoring scripts, which should enable further research in check-worthiness estimation.
AB - We present an overview of the CLEF-2018 CheckThat! Lab on Automatic Identification and Verification of Political Claims, with focus on Task 1: Check-Worthiness. The task asks to predict which claims in a political debate should be prioritized for fact-checking. In particular, given a debate or a political speech, the goal was to produce a ranked list of its sentences based on their worthiness for fact checking. We offered the task in both English and Arabic, based on debates from the 2016 US Presidential Campaign, as well as on some speeches during and after the campaign. A total of 30 teams registered to participate in the Lab and seven teams actually submitted systems for Task 1. The most successful approaches used by the participants relied on recurrent and multi-layer neural networks, as well as on combinations of distributional representations, on matchings claims' vocabulary against lexicons, and on measures of syntactic dependency. The best systems achieved mean average precision of 0.18 and 0.15 on the English and on the Arabic test datasets, respectively. This leaves large room for further improvement, and thus we release all datasets and the scoring scripts, which should enable further research in check-worthiness estimation.
KW - Check-worthiness
KW - Computational journalism
KW - Fact-checking
KW - Veracity
UR - http://www.scopus.com/inward/record.url?scp=85051071689&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85051071689
SN - 1613-0073
VL - 2125
JO - CEUR Workshop Proceedings
JF - CEUR Workshop Proceedings
Y2 - 10 September 2018 through 14 September 2018
ER -