Ring That Bell: A Corpus and Method for Multimodal Metaphor Detection in Videos

Khalid Alnajjar, Mika Hämäläinen, Shuo Zhang


Abstract
We present the first openly available multimodal metaphor annotated corpus. The corpus consists of videos including audio and subtitles that have been annotated by experts. Furthermore, we present a method for detecting metaphors in the new dataset based on the textual content of the videos. The method achieves a high F1-score (62%) for metaphorical labels. We also experiment with other modalities and multimodal methods; however, these methods did not out-perform the text-based model. In our error analysis, we do identify that there are cases where video could help in disambiguating metaphors, however, the visual cues are too subtle for our model to capture. The data is available on Zenodo.
Anthology ID:
2022.flp-1.4
Volume:
Proceedings of the 3rd Workshop on Figurative Language Processing (FLP)
Month:
December
Year:
2022
Address:
Abu Dhabi, United Arab Emirates (Hybrid)
Editors:
Debanjan Ghosh, Beata Beigman Klebanov, Smaranda Muresan, Anna Feldman, Soujanya Poria, Tuhin Chakrabarty
Venue:
Fig-Lang
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
24–33
Language:
URL:
https://aclanthology.org/2022.flp-1.4
DOI:
10.18653/v1/2022.flp-1.4
Bibkey:
Cite (ACL):
Khalid Alnajjar, Mika Hämäläinen, and Shuo Zhang. 2022. Ring That Bell: A Corpus and Method for Multimodal Metaphor Detection in Videos. In Proceedings of the 3rd Workshop on Figurative Language Processing (FLP), pages 24–33, Abu Dhabi, United Arab Emirates (Hybrid). Association for Computational Linguistics.
Cite (Informal):
Ring That Bell: A Corpus and Method for Multimodal Metaphor Detection in Videos (Alnajjar et al., Fig-Lang 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.flp-1.4.pdf
Video:
 https://aclanthology.org/2022.flp-1.4.mp4