Looking at the Unseen: Effective Sampling of Non-Related Propositions for Argument Mining

Ramon Ruiz-Dolz, Debela Gemechu, Zlata Kikteva, Chris Reed


Abstract
Traditionally, argument mining research has approached the task of automatic identification of argument structures by using existing definitions of what constitutes an argument, while leaving the equally important matter of what does not qualify as an argument unaddressed. With the ability to distinguish between what is and what is not a natural language argument being at the core of argument mining as a field, it is interesting that no previous work has explored approaches to effectively select non-related propositions (i.e., propositions that are not connected through an argumentative relation, such as support or attack) that improve the data for learning argument mining tasks better. In this paper, we address the question of how to effectively sample non-related propositions from six different argument mining corpora belonging to different domains and encompassing both monologue and dialogue forms of argumentation. To that end, in addition to considering undersampling baselines from previous work, we propose three new sampling strategies relying on context (i.e., short/long) and the semantic similarity between propositions. Our results indicate that using more informed sampling strategies improves the performance, not only when evaluating models on their respective test splits, but also in the case of cross-domain evaluation.
Anthology ID:
2025.coling-main.145
Volume:
Proceedings of the 31st International Conference on Computational Linguistics
Month:
January
Year:
2025
Address:
Abu Dhabi, UAE
Editors:
Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Venue:
COLING
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2131–2143
Language:
URL:
https://aclanthology.org/2025.coling-main.145/
DOI:
Bibkey:
Cite (ACL):
Ramon Ruiz-Dolz, Debela Gemechu, Zlata Kikteva, and Chris Reed. 2025. Looking at the Unseen: Effective Sampling of Non-Related Propositions for Argument Mining. In Proceedings of the 31st International Conference on Computational Linguistics, pages 2131–2143, Abu Dhabi, UAE. Association for Computational Linguistics.
Cite (Informal):
Looking at the Unseen: Effective Sampling of Non-Related Propositions for Argument Mining (Ruiz-Dolz et al., COLING 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.coling-main.145.pdf