Towards a Real-Time Multimodal Emotion Estimation Model for Dialogue Systems

Jingjing Jiang


Abstract
This position paper presents my research interest in establishing human-like chat-oriented dialogue systems. To this end, my work focuses on two main areas: the construction and utilization of multimodal datasets and real-time multimodal affective computing. I discuss the limitations of current multimodal dialogue corpora and multimodal affective computing models. As a solution, I have constructed a human-human dialogue dataset containing various synchronized multimodal information, and I have conducted preliminary analyses on it. In future work, I will further analyze the collected data and build a real-time multimodal emotion estimation model for dialogue systems.
Anthology ID:
2024.yrrsds-1.22
Volume:
Proceedings of the 20th Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems
Month:
September
Year:
2024
Address:
Kyoto, Japan
Editors:
Koji Inoue, Yahui Fu, Agnes Axelsson, Atsumoto Ohashi, Brielen Madureira, Yuki Zenimoto, Biswesh Mohapatra, Armand Stricker, Sopan Khosla
Venues:
YRRSDS | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
60–61
Language:
URL:
https://aclanthology.org/2024.yrrsds-1.22
DOI:
Bibkey:
Cite (ACL):
Jingjing Jiang. 2024. Towards a Real-Time Multimodal Emotion Estimation Model for Dialogue Systems. In Proceedings of the 20th Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems, pages 60–61, Kyoto, Japan. Association for Computational Linguistics.
Cite (Informal):
Towards a Real-Time Multimodal Emotion Estimation Model for Dialogue Systems (Jiang, YRRSDS-WS 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.yrrsds-1.22.pdf