Amanda: Adaptively Modality-Balanced Domain Adaptation for Multimodal Emotion Recognition

Xinxin Zhang, Jun Sun, Simin Hong, Taihao Li


Abstract
This paper investigates unsupervised multimodal domain adaptation for multimodal emotion recognition, which is a solution for data scarcity yet remains under studied. Due to the varying distribution discrepancies of different modalities between source and target domains, the primary challenge lies in how to balance the domain alignment across modalities to guarantee they are all well aligned. To achieve this, we first develop our model based on the information bottleneck theory to learn optimal representation for each modality independently. Then, we align the domains via matching the label distributions and the representations. In order to balance the representation alignment, we propose to minimize a surrogate of the alignment losses, which is equivalent to adaptively adjusting the weights of the modalities throughout training, thus achieving balanced domain alignment across modalities. Overall, the proposed approach features Adaptively modality-balanced domain adaptation, dubbed Amanda, for multimodal emotion recognition. Extensive empirical results on commonly used benchmark datasets demonstrate that Amanda significantly outperforms competing approaches. The code is available at https://github.com/sunjunaimer/Amanda.
Anthology ID:
2024.findings-acl.859
Volume:
Findings of the Association for Computational Linguistics ACL 2024
Month:
August
Year:
2024
Address:
Bangkok, Thailand and virtual meeting
Editors:
Lun-Wei Ku, Andre Martins, Vivek Srikumar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
14448–14458
Language:
URL:
https://aclanthology.org/2024.findings-acl.859
DOI:
Bibkey:
Cite (ACL):
Xinxin Zhang, Jun Sun, Simin Hong, and Taihao Li. 2024. Amanda: Adaptively Modality-Balanced Domain Adaptation for Multimodal Emotion Recognition. In Findings of the Association for Computational Linguistics ACL 2024, pages 14448–14458, Bangkok, Thailand and virtual meeting. Association for Computational Linguistics.
Cite (Informal):
Amanda: Adaptively Modality-Balanced Domain Adaptation for Multimodal Emotion Recognition (Zhang et al., Findings 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.findings-acl.859.pdf