A General Framework for Multimodal Argument Persuasiveness Classification of Tweets

Mohammad Soltani, Julia Romberg


Abstract
An important property of argumentation concerns the degree of its persuasiveness, which can be influenced by various modalities. On social media platforms, individuals usually have the option of supporting their textual statements with images. The goals of the ImageArg shared task, held with ArgMining 2023, were therefore (A) to classify tweet stances considering both modalities and (B) to predict the influence of an image on the persuasiveness of a tweet text. In this paper, we present our proposed methodology that shows strong performance on both tasks, placing 3rd team on the leaderboard in each case with F1 scores of 0.8273 (A) and 0.5281 (B). The framework relies on pre-trained models to extract text and image features, which are then fed into a task-specific classification model. Our experiments highlighted that the multimodal vision and language model CLIP holds a specific importance in the extraction of features, in particular for task (A).
Anthology ID:
2023.argmining-1.15
Volume:
Proceedings of the 10th Workshop on Argument Mining
Month:
December
Year:
2023
Address:
Singapore
Editors:
Milad Alshomary, Chung-Chi Chen, Smaranda Muresan, Joonsuk Park, Julia Romberg
Venues:
ArgMining | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
148–156
Language:
URL:
https://aclanthology.org/2023.argmining-1.15
DOI:
10.18653/v1/2023.argmining-1.15
Bibkey:
Cite (ACL):
Mohammad Soltani and Julia Romberg. 2023. A General Framework for Multimodal Argument Persuasiveness Classification of Tweets. In Proceedings of the 10th Workshop on Argument Mining, pages 148–156, Singapore. Association for Computational Linguistics.
Cite (Informal):
A General Framework for Multimodal Argument Persuasiveness Classification of Tweets (Soltani & Romberg, ArgMining-WS 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.argmining-1.15.pdf