Zhengzhong Tu
2026
A Survey on LLM-based Conversational User Simulation
Bo Ni | Yu Wang | Leyao Wang | Branislav Kveton | Franck Dernoncourt | Yu Xia | Hongjie Chen | Reuben Luera | Samyadeep Basu | Subhojyoti Mukherjee | Puneet Mathur | Nesreen K. Ahmed | Junda Wu | Li Li | Huixin Zhang | Ruiyi Zhang | Tong Yu | Sungchul Kim | Jiuxiang Gu | Zhengzhong Tu | Alexa Siu | Zichao Wang | Seunghyun Yoon | Nedim Lipka | Namyong Park | Zihao Lin | Trung Bui | Yue Zhao | Tyler Derr | Ryan A. Rossi
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Bo Ni | Yu Wang | Leyao Wang | Branislav Kveton | Franck Dernoncourt | Yu Xia | Hongjie Chen | Reuben Luera | Samyadeep Basu | Subhojyoti Mukherjee | Puneet Mathur | Nesreen K. Ahmed | Junda Wu | Li Li | Huixin Zhang | Ruiyi Zhang | Tong Yu | Sungchul Kim | Jiuxiang Gu | Zhengzhong Tu | Alexa Siu | Zichao Wang | Seunghyun Yoon | Nedim Lipka | Namyong Park | Zihao Lin | Trung Bui | Yue Zhao | Tyler Derr | Ryan A. Rossi
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
User simulation has long played a vital role in computer science due to its potential to support a wide range of applications. Language, as the primary medium of human communication, forms the foundation of social interaction and behavior. Consequently, simulating conversational behavior has become a key area of study. Recent advancements in large language models (LLMs) have significantly catalyzed progress in this domain by enabling high-fidelity generation of synthetic user conversation. In this paper, we survey recent advancements in LLM-based conversational user simulation. We introduce a novel taxonomy covering user granularity and simulation objectives. Additionally, we systematically analyze core techniques and evaluation methodologies. We aim to keep the research community informed of the latest advancements in conversational user simulation and to further facilitate future research by identifying open challenges and organizing existing work under a unified framework.
2025
Re-Align: Aligning Vision Language Models via Retrieval-Augmented Direct Preference Optimization
Shuo Xing | Peiran Li | Yuping Wang | Ruizheng Bai | Yueqi Wang | Chan-Wei Hu | Chengxuan Qian | Huaxiu Yao | Zhengzhong Tu
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Shuo Xing | Peiran Li | Yuping Wang | Ruizheng Bai | Yueqi Wang | Chan-Wei Hu | Chengxuan Qian | Huaxiu Yao | Zhengzhong Tu
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
The emergence of large Vision Language Models (VLMs) has broadened the scope and capabilities of single-modal Large Language Models (LLMs) by integrating visual modalities, thereby unlocking transformative cross-modal applications in a variety of real-world scenarios. Despite their impressive performance, VLMs are prone to significant hallucinations, particularly in the form of cross-modal inconsistencies. Building on the success of Reinforcement Learning from Human Feedback (RLHF) in aligning LLMs, recent advancements have focused on applying direct preference optimization (DPO) on carefully curated datasets to mitigate these issues. Yet, such approaches typically introduce preference signals in a brute-force manner, neglecting the crucial role of visual information in the alignment process. In this paper, we introduce Re-Align, a novel alignment framework that leverages image retrieval to construct a dual-preference dataset, effectively incorporating both textual and visual preference signals. We further introduce rDPO, an extension of the standard direct preference optimization that incorporates an additional visual preference objective during fine-tuning. Our experimental results demonstrate that Re-Align not only mitigates hallucinations more effectively than previous methods but also yields significant performance gains in general visual question-answering (VQA) tasks. Moreover, we show that Re-Align maintains robustness and scalability across a wide range of VLM sizes and architectures. This work represents a significant step forward in aligning multimodal LLMs, paving the way for more reliable and effective cross-modal applications.
Search
Fix author
Co-authors
- Nesreen K. Ahmed 1
- Ruizheng Bai 1
- Samyadeep Basu 1
- Trung Bui 1
- Hongjie Chen 1
- Franck Dernoncourt 1
- Tyler Derr 1
- Jiuxiang Gu 1
- Chan-Wei Hu 1
- Sungchul Kim 1
- Branislav Kveton 1
- Peiran Li 1
- Li Li 1
- Zihao Lin 1
- Nedim Lipka 1
- Reuben Luera 1
- Puneet Mathur 1
- Subhojyoti Mukherjee 1
- Bo Ni 1
- Namyong Park 1
- Chengxuan Qian 1
- Ryan A. Rossi 1
- Alexa Siu 1
- Yuping Wang 1
- Yueqi Wang 1
- Yu Wang 1
- Leyao Wang 1
- Zichao Wang 1
- Junda Wu 1
- Yu Xia 1
- Shuo Xing 1
- Huaxiu Yao 1
- Seunghyun Yoon 1
- Tong Yu 1
- Huixin Zhang 1
- Ruiyi Zhang 1
- Yue Zhao 1