Zhe-Yan Li
2025
Creativity in LLM-based Multi-Agent Systems: A Survey
Yi-Cheng Lin
|
Kang-Chieh Chen
|
Zhe-Yan Li
|
Tzu-Heng Wu
|
Tzu-Hsuan Wu
|
Kuan-Yu Chen
|
Hung-yi Lee
|
Yun-Nung Chen
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Large language model (LLM)-driven multi-agent systems (MAS) are transforming how humans and AIs collaboratively generate ideas and artifacts. While existing surveys provide comprehensive overviews of MAS infrastructures, they largely overlook the dimension of creativity, including how novel outputs are generated and evaluated, how creativity informs agent personas, and how creative workflows are coordinated. This is the first survey dedicated to creativity in MAS. We focus on text and image generation tasks, and present:(1) a taxonomy of agent proactivity and persona design;(2) an overview of generation techniques, including divergent exploration, iterative refinement, and collaborative synthesis, as well as relevant datasets and evaluation metrics; and(3) a discussion of key challenges, such as inconsistent evaluation standards, insufficient bias mitigation, coordination conflicts, and the lack of unified benchmarks.This survey offers a structured framework and roadmap for advancing the development, evaluation, and standardization of creative MAS.
Text-centric Alignment for Bridging Test-time Unseen Modality
Yun-Da Tsai
|
Ting-Yu Yen
|
Pei-Fu Guo
|
Zhe-Yan Li
|
Shou-De Lin
Findings of the Association for Computational Linguistics: EMNLP 2025
This paper addresses the challenge of handling unseen modalities and dynamic modality combinations at test time with our proposed text-centric alignment method. This training-free alignment approach unifies different input modalities into a single semantic text representation by leveraging in-context learning with Large Language Models and uni-modal foundation models. Our method significantly enhances the ability to manage unseen, diverse, and unpredictable modality combinations, making it suitable for both generative and discriminative models to adopt on top. Our extensive experiments primarily evaluate on discriminative tasks, demonstrating that our approach is essential for LLMs to achieve strong modality alignment performance. It also surpasses the limitations of traditional fixed-modality frameworks in embedding representations. This study contributes to the field by offering a flexible and effective solution for real-world applications where modality availability is dynamic and uncertain.
Search
Fix author
Co-authors
- Kang-Chieh Chen 1
- Kuan-Yu Chen 1
- Yun-Nung Chen 1
- Pei-Fu Guo 1
- Hung-Yi Lee 1
- show all...