Yifan Liu
Papers on this page may belong to the following people: Yifan Liu, Yifan Simon Liu
2025
AssoCiAm: A Benchmark for Evaluating Association Thinking while Circumventing Ambiguity
Yifan Liu | Wenkuan Zhao | Shanshan Zhong | Jinghui Qin | Mingfu Liang | Zhongzhan Huang | Wushao Wen
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Yifan Liu | Wenkuan Zhao | Shanshan Zhong | Jinghui Qin | Mingfu Liang | Zhongzhan Huang | Wushao Wen
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Recent advancements in multimodal large language models (MLLMs) have garnered significant attention, offering a promising pathway toward artificial general intelligence (AGI). Among the essential capabilities required for AGI, creativity has emerged as a critical trait for MLLMs, with association serving as its foundation. Association reflects a model’s ability to think creatively, making it vital to evaluate and understand. While several frameworks have been proposed to assess associative ability, they often overlook the inherent ambiguity in association tasks, which arises from the divergent nature of associations and undermines the reliability of evaluations. To address this issue, we decompose ambiguity into two types—internal ambiguity and external ambiguity—and introduce AssoCiAm, a benchmark designed to evaluate associative ability while circumventing the ambiguity through a hybrid computational method. We then conduct extensive experiments on MLLMs, revealing a strong positive correlation between cognition and association. Additionally, we observe that the presence of ambiguity in the evaluation process causes MLLMs’ behavior to become more random-like. Finally, we validate the effectiveness of our method in ensuring more accurate and reliable evaluations. See Project Page for the data and codes.
User Feedback Alignment for LLM-powered Exploration in Large-scale Recommendation Systems
Jianling Wang | Yifan Liu | Yinghao Sun | Xuejian Ma | Yueqi Wang | He Ma | Zhengyang Su | Minmin Chen | Mingyan Gao | Onkar Dalal | Ed H. Chi | Lichan Hong | Ningren Han | Haokai Lu
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 6: Industry Track)
Jianling Wang | Yifan Liu | Yinghao Sun | Xuejian Ma | Yueqi Wang | He Ma | Zhengyang Su | Minmin Chen | Mingyan Gao | Onkar Dalal | Ed H. Chi | Lichan Hong | Ningren Han | Haokai Lu
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 6: Industry Track)
Exploration, the act of broadening user experiences beyond their established preferences, is challenging in large-scale recommendation systems due to feedback loops and limited signals on user exploration patterns. Large Language Models (LLMs) offer potential solutions by leveraging their world knowledge to recommend novel content outside these loops. A key challenge is aligning LLMs with user preferences while preserving their knowledge and reasoning. To enhance planning for new user interests using LLMs, this paper introduces a novel approach that combines hierarchical planning with LLM inference-time scaling. This method aims to improve recommendation relevancy without compromising novelty. We decouple novelty and user-alignment, training separate LLMs for each objective. We then scale up the novelty-focused LLM’s inference and select the best-of-n predictions using the user-aligned LLM. Live experiments demonstrate efficacy, showing significant gains in both user satisfaction (measured by watch activity and active user counts) and exploration diversity.
2024
Retrieval Augmented Fact Verification by Synthesizing Contrastive Arguments
Zhenrui Yue | Huimin Zeng | Lanyu Shang | Yifan Liu | Yang Zhang | Dong Wang
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Zhenrui Yue | Huimin Zeng | Lanyu Shang | Yifan Liu | Yang Zhang | Dong Wang
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
The rapid propagation of misinformation poses substantial risks to public interest. To combat misinformation, large language models (LLMs) are adapted to automatically verify claim credibility. Nevertheless, existing methods heavily rely on the embedded knowledge within LLMs and / or black-box APIs for evidence collection, leading to subpar performance with smaller LLMs or upon unreliable context. In this paper, we propose retrieval augmented fact verification through the synthesis of contrasting arguments (RAFTS). Upon input claims, RAFTS starts with evidence retrieval, where we design a retrieval pipeline to collect and re-rank relevant documents from verifiable sources. Then, RAFTS forms contrastive arguments (i.e., supporting or refuting) conditioned on the retrieved evidence. In addition, RAFTS leverages an embedding model to identify informative demonstrations, followed by in-context prompting to generate the prediction and explanation. Our method effectively retrieves relevant documents as evidence and evaluates arguments from varying perspectives, incorporating nuanced information for fine-grained decision-making. Combined with informative in-context examples as prior, RAFTS achieves significant improvements to supervised and LLM baselines without complex prompts. We demonstrate the effectiveness of our method through extensive experiments, where RAFTS can outperform GPT-based methods with a significantly smaller 7B LLM.
2023
Multi-Stage Coarse-to-Fine Contrastive Learning for Conversation Intent Induction
Caiyuan Chu | Ya Li | Yifan Liu | Jia-Chen Gu | Quan Liu | Yongxin Ge | Guoping Hu
Proceedings of the Eleventh Dialog System Technology Challenge
Caiyuan Chu | Ya Li | Yifan Liu | Jia-Chen Gu | Quan Liu | Yongxin Ge | Guoping Hu
Proceedings of the Eleventh Dialog System Technology Challenge
Intent recognition is critical for task-oriented dialogue systems. However, for emerging domains and new services, it is difficult to accurately identify the key intent of a conversation due to time-consuming data annotation and comparatively poor model transferability. Therefore, the automatic induction of dialogue intention is very important for intelligent dialogue systems. This paper presents our solution to Track 2 of Intent Induction from Conversations for Task-Oriented Dialogue at the Eleventh Dialogue System Technology Challenge (DSTC11). The essence of intention clustering lies in distinguishing the representation of different dialogue utterances. The key to automatic intention induction is that, for any given set of new data, the sentence representation obtained by the model can be well distinguished from different labels. Therefore, we propose a multi-stage coarse-to-fine contrastive learning model training scheme including unsupervised contrastive learning pre-training, supervised contrastive learning pre-training, and fine-tuning with joint contrastive learning and clustering to obtain a better dialogue utterance representation model for the clustering task. In the released DSTC11 Track 2 evaluation results, our proposed system ranked first on both of the two subtasks of this Track.
2019
AiFu at SemEval-2019 Task 10: A Symbolic and Sub-symbolic Integrated System for SAT Math Question Answering
Yifan Liu | Keyu Ding | Yi Zhou
Proceedings of the 13th International Workshop on Semantic Evaluation
Yifan Liu | Keyu Ding | Yi Zhou
Proceedings of the 13th International Workshop on Semantic Evaluation
AiFu has won the first place in the SemEval-2019 Task 10 - ”Math Question Answering”competition. This paper is to describe how it works technically and to report and analyze some essential experimental results
Search
Fix author
Co-authors
- Minmin Chen 1
- Ed H. Chi 1
- Caiyuan Chu 1
- Onkar Dalal 1
- Keyu Ding 1
- Mingyan Gao 1
- Yongxin Ge 1
- Jia-Chen Gu 1
- Ningren Han 1
- Lichan Hong 1
- Guoping Hu 1
- Zhongzhan Huang 1
- Ya Li 1
- Mingfu Liang 1
- Quan Liu 1
- Haokai Lu 1
- Xuejian Ma 1
- He Ma 1
- Jinghui Qin 1
- Lanyu Shang 1
- Zhengyang Su 1
- Yinghao Sun 1
- Dong Wang 1
- Jianling Wang 1
- Yueqi Wang 1
- Wushao Wen 1
- Zhenrui Yue 1
- Huimin Zeng 1
- Yang Zhang 1
- Wenkuan Zhao 1
- Shanshan Zhong 1
- Yi Zhou 1