Multimodal Agentic Dialogue Systems for Situated Human-Robot Interaction

Virgile Sucal


Abstract
This position paper presents the integration of dialogue systems into situated robotics, emphasizing the use of contextual information—particularly audiovisual perceptions—to inform dialogue policies. A central objective is the development of interaction policies that dynamically select contextually appropriate actions aligned with the user’s intentions and needs. The works presented in this paper explore proactive decision-making mechanisms in multimodal interaction settings and seek to enhance robotic expressiveness through nonverbal communication cues. Current efforts focus on evaluating and comparing approaches such as agentic workflows and reinforcement learning within a unified framework, aiming to facilitate more consistent and contextually aware human–robot interaction.
Anthology ID:
2025.yrrsds-1.8
Volume:
Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems
Month:
August
Year:
2025
Address:
Avignon, France
Editors:
Ryan Whetten, Virgile Sucal, Anh Ngo, Kranti Chalamalasetti, Koji Inoue, Gaetano Cimino, Zachary Yang, Yuki Zenimoto, Ricardo Rodriguez
Venue:
YRRSDS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
20–24
Language:
URL:
https://aclanthology.org/2025.yrrsds-1.8/
DOI:
Bibkey:
Cite (ACL):
Virgile Sucal. 2025. Multimodal Agentic Dialogue Systems for Situated Human-Robot Interaction. In Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems, pages 20–24, Avignon, France. Association for Computational Linguistics.
Cite (Informal):
Multimodal Agentic Dialogue Systems for Situated Human-Robot Interaction (Sucal, YRRSDS 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.yrrsds-1.8.pdf