Qingyun Wu
2026
Do Images Speak Louder than Words? Investigating the Effect of Textual Misinformation in VLMs
Chi Zhang | Wenxuan Ding | Jiale Liu | Mingrui Wu | Qingyun Wu | Ray Mooney
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Chi Zhang | Wenxuan Ding | Jiale Liu | Mingrui Wu | Qingyun Wu | Ray Mooney
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Vision-Language Models (VLMs) have shown strong multimodal reasoning capability on Visual-Question-Answering (VQA) benchmarks. However, their robustness against textual misinformation remains under-explored. While existing research has extensively studied the effect of misinformation in text-only domains, it is not clear how VLMs arbitrate between contradictory information from different modalities. To bridge the gap, we first propose the ConText-VQA (i.e. Conflicting Text) dataset, consisting of image-question pairs together with systematically generated persuasive prompts that deliberately conflict with visual evidence. Then, a thorough testing framework is designed and executed to benchmark the susceptibility of various models to these conflicting textual inputs. Comprehensive experiments over 11 state-of-the-art VLMs reveal that these models are indeed vulnerable to misleading prompts, often overriding clear visual evidence in favor of the conflicting text, and show an average performance drop of over 48.2% after only one round of persuasive conversation. Our findings highlight a critical limitation in current VLMs and underscore the need for improved robustness against textual manipulation.
2025
SimpleDoc: Multi‐Modal Document Understanding with Dual‐Cue Page Retrieval and Iterative Refinement
Chelsi Jain | Yiran Wu | Yifan Zeng | Jiale Liu | Shengyu Dai | Zhenwen Shao | Qingyun Wu | Huazheng Wang
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Chelsi Jain | Yiran Wu | Yifan Zeng | Jiale Liu | Shengyu Dai | Zhenwen Shao | Qingyun Wu | Huazheng Wang
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Document Visual Question Answering (DocVQA) is a practical yet challenging task, which is to ask questions based on documents while referring to multiple pages and different modalities of information, e.g., images and tables. To handle multi-modality, recent methods follow a similar Retrieval Augmented Generation (RAG) pipeline, but utilize Visual Language Models (VLMs) based embedding model to embed and retrieve relevant pages as images, and generate answers with VLMs that can accept an image as input. In this paper, we introduce SimpleDoc, a lightweight yet powerful retrieval - augmented framework for DocVQA. It boosts evidence page gathering by first retrieving candidates through embedding similarity and then filtering and re-ranking these candidates based on page summaries. A single VLM-based reasoner agent repeatedly invokes this dual-cue retriever, iteratively pulling fresh pages into a working memory until the question is confidently answered. SimpleDoc outperforms previous baselines by 3.2% on average on 4 DocVQA datasets with much fewer pages retrieved. Our code is available at https://github.com/ag2ai/SimpleDoc.
Divide, Optimize, Merge: Scalable Fine-Grained Generative Optimization for LLM Agents
Jiale Liu | Yifan Zeng | Shaokun Zhang | Chi Zhang | Malte Højmark-Bertelsen | Marie Normann Gadeberg | Huazheng Wang | Qingyun Wu
Findings of the Association for Computational Linguistics: EMNLP 2025
Jiale Liu | Yifan Zeng | Shaokun Zhang | Chi Zhang | Malte Højmark-Bertelsen | Marie Normann Gadeberg | Huazheng Wang | Qingyun Wu
Findings of the Association for Computational Linguistics: EMNLP 2025
LLM-based optimization has shown remarkable potential in improving agentic systems. However, the conventional approach of prompting LLM-based generative optimizer with the trajectories on the whole training dataset in a single pass becomes untenable as datasets grow, leading to context window overflow and degraded pattern recognition. To address these challenges, we propose Fine-grained Generative Optimization (FGO), a scalable framework that divides large optimization tasks into manageable subsets, performs targeted optimizations, and systematically combines optimized components through progressive merging.Evaluation across ALFWorld, LogisticsQA, and GAIA benchmarks demonstrates that FGO outperforms conventional approach by 1.6-8.6% while reducing average prompt token consumption by 56.3%. Our framework provides a practical solution for scaling up LLM-based generative optimization of increasingly sophisticated agentic systems. Further analysis demonstrates that FGO achieves the most consistent performance gain in all training dataset sizes, showcasing its scalability and efficiency.
2024
Assessing and Verifying Task Utility in LLM-Powered Applications
Negar Arabzadeh | Siqing Huo | Nikhil Mehta | Qingyun Wu | Chi Wang | Ahmed Hassan Awadallah | Charles L. A. Clarke | Julia Kiseleva
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Negar Arabzadeh | Siqing Huo | Nikhil Mehta | Qingyun Wu | Chi Wang | Ahmed Hassan Awadallah | Charles L. A. Clarke | Julia Kiseleva
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
The rapid development of Large Language Models (LLMs) has led to a surge in applications that facilitate collaboration among multiple agents, assisting humans in their daily tasks. However, a significant gap remains in assessing to what extent LLM-powered applications genuinely enhance user experience and task execution efficiency. This highlights the need to verify utility of LLM-powered applications, particularly by ensuring alignment between the application’s functionality and end-user needs. We introduce AgentEval, a novel framework designed to simplify the utility verification process by automatically proposing a set of criteria tailored to the unique purpose of any given application. This allows for a comprehensive assessment, quantifying the utility of an application against the suggested criteria. We present a comprehensive analysis of the effectiveness and robustness of AgentEval for two open source datasets including Math Problem solving and ALFWorld House-hold related tasks. For reproducibility purposes, we make the data, code and all the logs publicly available at https://github.com/Narabzad/AgentEval