VeryfIT - Benchmark of Fact-Checked Claims for Italian: A CALAMITA Challenge

Jacopo Gili, Viviana Patti, Lucia Passaro, Tommaso Caselli


Abstract
Achieving factual accuracy is a known pending issue for language models. Their design centered around the interactive component of user interaction and the extensive use of “spontaneous” training data, has made them highly adept at conversational tasks but not fully reliable in terms of factual correctness. VeryfIT addresses this issue by evaluating the in-memory factual knowledge of language models on data written by professional fact-checkers, posing it as a true or false question.Topics of the statements vary but most are in specific domains related to the Italian government, policies, and social issues. The task presents several challenges: extracting statements from segments of speeches, determining appropriate contextual relevance both temporally and factually, and ultimately verifying the accuracy of the statements.
Anthology ID:
2024.clicit-1.123
Volume:
Proceedings of the 10th Italian Conference on Computational Linguistics (CLiC-it 2024)
Month:
December
Year:
2024
Address:
Pisa, Italy
Editors:
Felice Dell'Orletta, Alessandro Lenci, Simonetta Montemagni, Rachele Sprugnoli
Venue:
CLiC-it
SIG:
Publisher:
CEUR Workshop Proceedings
Note:
Pages:
1116–1124
Language:
URL:
https://aclanthology.org/2024.clicit-1.123/
DOI:
Bibkey:
Cite (ACL):
Jacopo Gili, Viviana Patti, Lucia Passaro, and Tommaso Caselli. 2024. VeryfIT - Benchmark of Fact-Checked Claims for Italian: A CALAMITA Challenge. In Proceedings of the 10th Italian Conference on Computational Linguistics (CLiC-it 2024), pages 1116–1124, Pisa, Italy. CEUR Workshop Proceedings.
Cite (Informal):
VeryfIT - Benchmark of Fact-Checked Claims for Italian: A CALAMITA Challenge (Gili et al., CLiC-it 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.clicit-1.123.pdf