Lin Lee Cheong
2026
Learning to Ideate for Machine Learning Engineering Agents
Yunxiang Zhang | Kang Zhou | Zhichao Xu | Kiran Ramnath | Yun Zhou | Sangmin Woo | Haibo Ding | Lin Lee Cheong
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers)
Yunxiang Zhang | Kang Zhou | Zhichao Xu | Kiran Ramnath | Yun Zhou | Sangmin Woo | Haibo Ding | Lin Lee Cheong
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers)
Existing machine learning engineering (MLE) agents struggle to iteratively optimize their implemented algorithms for effectiveness. To address this, we introduce MLE-Ideator, a dual-agent framework that separates ideation from implementation. In our system, an implementation agent can request strategic help from a dedicated Ideator. We show this approach is effective in two ways. First, in a training-free setup, our framework significantly outperforms implementation-only agent baselines on MLE-Bench. Second, we demonstrate that the Ideator can be trained with reinforcement learning (RL) to generate more effective ideas. With only 1K training samples from 10 MLE tasks, our RL-trained Qwen3-8B Ideator achieves an 11.5% relative improvement compared to its untrained counterpart and surpasses Claude Sonnet 3.5. These results highlights a promising path toward training strategic AI systems for scientific discovery.
SALT: Step-level Advantage Assignment for Long-horizon Agents via Trajectory Graph
Jiazheng Li | Yawei Wang | Qiaojing Yan | Yijun Tian | Zhichao Xu | Huan Song | Panpan Xu | Lin Lee Cheong
Findings of the Association for Computational Linguistics: EACL 2026
Jiazheng Li | Yawei Wang | Qiaojing Yan | Yijun Tian | Zhichao Xu | Huan Song | Panpan Xu | Lin Lee Cheong
Findings of the Association for Computational Linguistics: EACL 2026
Large Language Models (LLMs) have demonstrated remarkable capabilities, enabling language agents to excel at single-turn tasks. However, their application to complex, multi-step, and long-horizon tasks remains challenging. While reinforcement learning (RL) offers a promising avenue for addressing these challenges, mainstream approaches typically rely solely on sparse, outcome-based rewards — a limitation that becomes especially problematic for group-based RL algorithms lacking critic models, such as Group Relative Policy Optimization (GRPO). In such methods, uniformly rewarding or penalizing all actions within a trajectory can lead to training instability and suboptimal policies, because beneficial and detrimental actions are often entangled across multi-step interactions. To address this challenge, we propose SALT, a novel and lightweight framework that provides a finer-grained advantage assignment, derived solely from outcome rewards. We achieve this by constructing a graph from trajectories of the same prompt, which allows us to quantify the quality of each step and assign advantages accordingly. Crucially, SALT is designed as a plug-and-play module that seamlessly integrates with existing group-based RL algorithms — requiring no modifications to the rollout procedure and introducing negligible computational overhead. Extensive experiments on the WebShop, ALFWorld, and AppWorld benchmarks with various model sizes demonstrate that SALT consistently improves performance. We also conduct a thorough analysis to validate the design choices behind SALT and offer actionable insights.
PromptPrism: A Linguistically-Inspired Taxonomy for Prompts
Sullam Jeoung | Yueyan Chen | Yi Zhang | Shuai Wang | Haibo Ding | Lin Lee Cheong
Findings of the Association for Computational Linguistics: EACL 2026
Sullam Jeoung | Yueyan Chen | Yi Zhang | Shuai Wang | Haibo Ding | Lin Lee Cheong
Findings of the Association for Computational Linguistics: EACL 2026
Prompts are the interface for eliciting the capabilities of large language models (LLMs). Understanding their structure and components is critical for analyzing LLM behavior and optimizing performance. However, the field lacks a comprehensive framework for systematic prompt analysis and understanding. We introduce PromptPrism, a linguistically-inspired taxonomy that enables prompt analysis across three hierarchical levels: functional structure, semantic component, and syntactic pattern. By applying linguistic concepts to prompt analysis, PromptPrism bridges traditional language understanding and modern LLM research, offering insights that purely empirical approaches might miss. We show the practical utility of PromptPrism by applying it to three applications: (1) a taxonomy-guided prompt refinement approach that automatically improves prompt quality and enhances model performance across a range of tasks; (2) a multi-dimensional dataset profiling method that extracts and aggregates structural, semantic, and syntactic characteristics from prompt datasets, enabling comprehensive analysis of prompt distributions and patterns; (3) a controlled experimental framework for prompt sensitivity analysis by quantifying the impact of semantic reordering and delimiter modifications on LLM performance. Our experimental results validate the effectiveness of our taxonomy across these applications, demonstrating that PromptPrism provides a foundation for refining, profiling, and analyzing prompts.
Diffusion Language Model Inference with Monte Carlo Tree Search
Zheng Huang | Kiran Ramnath | Yueyan Chen | Aosong Feng | Sangmin Woo | Balasubramaniam Srinivasan | Zhichao Xu | Kang Zhou | Shuai Wang | Haibo Ding | Lin Lee Cheong
Findings of the Association for Computational Linguistics: EACL 2026
Zheng Huang | Kiran Ramnath | Yueyan Chen | Aosong Feng | Sangmin Woo | Balasubramaniam Srinivasan | Zhichao Xu | Kang Zhou | Shuai Wang | Haibo Ding | Lin Lee Cheong
Findings of the Association for Computational Linguistics: EACL 2026
Diffusion language models (DLMs) have recently emerged as a compelling alternative to autoregressive generation, offering parallel generation and improved global coherence. During inference, DLMs generate text by iteratively denoising masked sequences in parallel; however, determining which positions to unmask and which tokens to commit forms a large combinatorial search problem. Existing inference methods approximate this search using heuristics, which often yield suboptimal decoding paths; other approaches instead rely on additional training to guide token selection. To introduce a principled search mechanism for DLMs inference, we introduce MEDAL, an inference-time scaling framework that integrates Monte Carlo Tree SEarch initialization for Diffusion LAnguage Model inference. We employ Monte Carlo Tree Search at the initialization stage to explore promising unmasking trajectories, providing a robust starting point for subsequent refinement. This design enables efficient inference-time scaling, allowing generation quality to improve as the search budget increases, without additional training. Across multiple benchmarks, MEDAL achieves up to 22.0% improvement over existing inference strategies, establishing a new paradigm for search-based inference in DLMs.
2025
IPR: Intelligent Prompt Routing with User-Controlled Quality-Cost Trade-offs
Aosong Feng | Balasubramaniam Srinivasan | Yun Zhou | Zhichao Xu | Kang Zhou | Sheng Guan | Yueyan Chen | Xian Wu | Ninad Kulkarni | Yi Zhang | Zhengyuan Shen | Dmitriy Bespalov | Soumya Smruti Mishra | Yifei Teng | Darren Yow-Bang Wang | Haibo Ding | Lin Lee Cheong
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Aosong Feng | Balasubramaniam Srinivasan | Yun Zhou | Zhichao Xu | Kang Zhou | Sheng Guan | Yueyan Chen | Xian Wu | Ninad Kulkarni | Yi Zhang | Zhengyuan Shen | Dmitriy Bespalov | Soumya Smruti Mishra | Yifei Teng | Darren Yow-Bang Wang | Haibo Ding | Lin Lee Cheong
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Routing incoming queries to the most cost-effective LLM while maintaining response quality poses a fundamental challenge in optimizing performance-cost trade-offs for large-scale commercial systems.We present IPR—a quality-constrained Intelligent Prompt Routing framework that dynamically selects optimal models based on predicted response quality and user-specified tolerance levels.IPR introduces three key innovations: (1) a modular architecture with lightweight quality estimators trained on 1.5M prompts annotated with calibrated quality scores, enabling fine-grained quality prediction across model families; (2) a user-controlled routing mechanism with tolerance parameter 𝜏 ∈ [0,1] that provides explicit control over quality-cost trade-offs; and (3) an extensible design using frozen encoders with model-specific adapters, reducing new model integration from days to hours. To rigorously train and evaluate IPR, we curate an industrial-level IPR dataset, a comprehensive benchmark containing 1.5 million examples with response quality annotations across 11 LLM candidates.Deployed on a major cloud platform, IPR achieves 43.9% cost reduction while maintaining quality parity with the strongest model in the Claude family and processes requests with sub-150ms latency.
A Systematic Survey of Automatic Prompt Optimization Techniques
Kiran Ramnath | Kang Zhou | Sheng Guan | Soumya Smruti Mishra | Xuan Qi | Zhengyuan Shen | Shuai Wang | Sangmin Woo | Sullam Jeoung | Yawei Wang | Haozhu Wang | Han Ding | Yuzhe Lu | Zhichao Xu | Yun Zhou | Balasubramaniam Srinivasan | Qiaojing Yan | Yueyan Chen | Haibo Ding | Panpan Xu | Lin Lee Cheong
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Kiran Ramnath | Kang Zhou | Sheng Guan | Soumya Smruti Mishra | Xuan Qi | Zhengyuan Shen | Shuai Wang | Sangmin Woo | Sullam Jeoung | Yawei Wang | Haozhu Wang | Han Ding | Yuzhe Lu | Zhichao Xu | Yun Zhou | Balasubramaniam Srinivasan | Qiaojing Yan | Yueyan Chen | Haibo Ding | Panpan Xu | Lin Lee Cheong
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Since the advent of large language models (LLMs), prompt engineering has been a crucial step for eliciting desired responses for various Natural Language Processing (NLP) tasks. However, prompt engineering remains an impediment for end users due to rapid advances in models, tasks, and associated best practices. To mitigate this, Automatic Prompt Optimization (APO) techniques have recently emerged that use various automated techniques to help improve the performance of LLMs on various tasks. In this paper, we present a comprehensive survey summarizing the current progress and remaining challenges in this field. We provide a formal definition of APO, a 5-part unifying framework, and then proceed to rigorously categorize all relevant works based on their salient features therein. We hope to spur further research guided by our framework.
CSPLADE: Learned Sparse Retrieval with Causal Language Models
Zhichao Xu | Aosong Feng | Yijun Tian | Haibo Ding | Lin Lee Cheong
Proceedings of the 14th International Joint Conference on Natural Language Processing and the 4th Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics
Zhichao Xu | Aosong Feng | Yijun Tian | Haibo Ding | Lin Lee Cheong
Proceedings of the 14th International Joint Conference on Natural Language Processing and the 4th Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics
In recent years, dense retrieval has been the focus of information retrieval (IR) research. While effective, dense retrieval produces uninterpretable dense vectors, and suffers from the drawback of large index size. Learned sparse retrieval (LSR) has emerged as promising alternative, achieving competitive retrieval performance while also being able to leverage the classical inverted index data structure for efficient retrieval. However, limited works have explored scaling LSR beyond BERT scale. In this work, we identify two challenges in training large language models (LLM) for LSR: (1) training instability during the early stage of contrastive training; (2) suboptimal performance due to pre-trained LLM’s unidirectional attention. To address these challenges, we propose two corresponding techniques: (1) a lightweight adaptation training phase to eliminate training instability; (2) two model variants to enable bidirectional information. With these techniques, we are able to train LSR models with 8B scale LLM, and achieve competitive retrieval performance with reduced index size. Furthermore, we are among the first to analyze the performance-efficiency tradeoff of LLM-based LSR model through the lens of model quantization. Our findings provide insights into adapting LLMs for efficient retrieval modeling.
Black-Box Visual Prompt Engineering for Mitigating Object Hallucination in Large Vision Language Models
Sangmin Woo | Kang Zhou | Yun Zhou | Shuai Wang | Sheng Guan | Haibo Ding | Lin Lee Cheong
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers)
Sangmin Woo | Kang Zhou | Yun Zhou | Shuai Wang | Sheng Guan | Haibo Ding | Lin Lee Cheong
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers)
Large Vision Language Models (LVLMs) often suffer from object hallucination, which undermines their reliability. Surprisingly, we find that simple object-based visual prompting—overlaying visual cues (e.g., bounding box, circle) on images—can significantly mitigate such hallucination; however, different visual prompts (VPs) vary in effectiveness. To address this, we propose Black-Box Visual Prompt Engineering (BBVPE), a framework to identify optimal VPs that enhance LVLM responses without needing access to model internals. Our approach employs a pool of candidate VPs and trains a router model to dynamically select the most effective VP for a given input image. This black-box approach is model-agnostic, making it applicable to both open-source and proprietary LVLMs. Evaluations on benchmarks such as POPE and CHAIR demonstrate that BBVPE effectively reduces object hallucination.
Search
Fix author
Co-authors
- Haibo Ding 7
- Zhichao Xu 6
- Kang Zhou 5
- Yueyan Chen 4
- Shuai Wang 4
- Sangmin Woo 4
- Yun Zhou 4
- Aosong Feng 3
- Sheng Guan 3
- Kiran Ramnath 3
- Balasubramaniam Srinivasan 3
- Sullam Jeoung 2
- Soumya Smruti Mishra 2
- Zhengyuan Shen 2
- Yijun Tian 2
- Yawei Wang 2
- Panpan Xu 2
- Qiaojing Yan 2
- Yi Zhang 2
- Dmitriy Bespalov 1
- Han Ding 1
- Zheng Huang 1
- Ninad Kulkarni 1
- Jiazheng Li 1
- Yuzhe Lu 1
- Xuan Qi 1
- Huan Song 1
- Yifei Teng 1
- Darren Yow-Bang Wang 1
- Haozhu Wang 1
- Xian Wu 1
- Yunxiang Zhang 1