Junyoung Youn
2024
Stark: Social Long-Term Multi-Modal Conversation with Persona Commonsense Knowledge
Young-Jun Lee
|
Dokyong Lee
|
Junyoung Youn
|
Kyeong-Jin Oh
|
Byungsoo Ko
|
Jonghwan Hyeon
|
Ho-Jin Choi
Findings of the Association for Computational Linguistics: EMNLP 2024
Humans share a wide variety of images related to their personal experiences within conversations via instant messaging tools. However, existing works focus on (1) image-sharing behavior in singular sessions, leading to limited long-term social interaction, and (2) a lack of personalized image-sharing behavior. In this work, we introduce , a large-scale long-term multi-modal dialogue dataset that covers a wide range of social personas in a multi-modality format, time intervals, and images. To construct automatically, we propose a novel multi-modal contextualization framework, , that generates long-term multi-modal dialogue distilled from ChatGPT and our proposed image aligner. Using our , we train a multi-modal conversation model, 7B, which demonstrates impressive visual imagination ability. Furthermore, we demonstrate the effectiveness of our dataset in human evaluation. The code, dataset, and model will be publicly released after publication.
Search
Co-authors
- Young-Jun Lee 1
- Dokyong Lee 1
- Kyeong-Jin Oh 1
- Byungsoo Ko 1
- Jonghwan Hyeon 1
- show all...