Continuous Attentive Multimodal Prompt Tuning for Few-Shot Multimodal Sarcasm Detection

Soumyadeep Jana, Animesh Dey, Ranbir Singh Sanasam


Abstract
With the steep rise in multimodal content on social media, multimodal sarcasm detection has gained widespread attention from research communities. Existing studies depend on large-scale data, which is challenging to obtain and expensive to annotate. Thus, investigating this problem in a few-shot scenario is required. Overtly complex multimodal models are prone to overfitting on in-domain data, which hampers their performance on out-of-distribution (OOD) data. To address these issues, we propose Continuous Attentive Multimodal Prompt Tuning model (CAMP), that leverages the prompt tuning paradigm to handle few-shot multimodal sarcasm detection. To overcome the siloed learning process of continuous prompt tokens, we design a novel, continuous multimodal attentive prompt where the continuous tokens intricately engage with both image and text tokens, enabling the assimilation of knowledge from different input modalities. Experimental results indicate that our method outperforms other multimodal baseline methods in the few-shot setting and OOD scenarios.
Anthology ID:
2024.conll-1.25
Volume:
Proceedings of the 28th Conference on Computational Natural Language Learning
Month:
November
Year:
2024
Address:
Miami, FL, USA
Editors:
Libby Barak, Malihe Alikhani
Venue:
CoNLL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
314–326
Language:
URL:
https://aclanthology.org/2024.conll-1.25
DOI:
Bibkey:
Cite (ACL):
Soumyadeep Jana, Animesh Dey, and Ranbir Singh Sanasam. 2024. Continuous Attentive Multimodal Prompt Tuning for Few-Shot Multimodal Sarcasm Detection. In Proceedings of the 28th Conference on Computational Natural Language Learning, pages 314–326, Miami, FL, USA. Association for Computational Linguistics.
Cite (Informal):
Continuous Attentive Multimodal Prompt Tuning for Few-Shot Multimodal Sarcasm Detection (Jana et al., CoNLL 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.conll-1.25.pdf