Zheyu Zhao
2025
Bridging Modality Gap for Effective Multimodal Sentiment Analysis in Fashion-related Social Media
Zheyu Zhao
|
Zhongqing Wang
|
Shichen Li
|
Hongling Wang
|
Guodong Zhou
Proceedings of the 31st International Conference on Computational Linguistics
Multimodal sentiment analysis for fashion-related social media is essential for understanding how consumers appraise fashion products across platforms like Instagram and Twitter, where both textual and visual elements contribute to sentiment expression. However, a notable challenge in this task is the modality gap, where the different information density between text and images hinders effective sentiment analysis. In this paper, we propose a novel multimodal framework that addresses this challenge by introducing pseudo data generated by a two-stage framework. We further utilize a multimodal fusion approach that efficiently integrates the information from various modalities for sentiment classification of fashion posts. Experiments conducted on a comprehensive dataset demonstrate that our framework significantly outperforms existing unimodal and multimodal baselines, highlighting its effectiveness in bridging the modality gap for more accurate sentiment classification in fashion-related social media posts.