Ke Wang
Renmin
Other people with similar names: Ke Wang
Unverified author pages with similar names: Ke Wang
2025
RACQC: Advanced Retrieval-Augmented Generation for Chinese Query Correction
Jinbo Su | Lingzhe Gao | Wei Li | Shihao Liu | Haojie Lei | Xinyi Wang | Yuanzhao Guo | Ke Wang | Daiting Shi | Dawei Yin
Findings of the Association for Computational Linguistics: EMNLP 2025
Jinbo Su | Lingzhe Gao | Wei Li | Shihao Liu | Haojie Lei | Xinyi Wang | Yuanzhao Guo | Ke Wang | Daiting Shi | Dawei Yin
Findings of the Association for Computational Linguistics: EMNLP 2025
In web search scenarios, erroneous queries frequently degrade users’ experience through irrelevant results, underscoring the pivotal role of Chinese Spelling Check (CSC) systems. Although large language models (LLMs) exhibit remarkable capabilities across many tasks, they face critical challenges in the CSC scenario: (1) poor generalization to rare entities in open-domain searches, and (2) failure to adapt to temporal entity variations due to static parameters, resulting in serious over-correction issues. To tackle this, we present RACQC, a Chinese Query Correction system with Retrieval-Augmented Generation (RAG) and multi-task learning. Specifically, our approach (1) integrates dynamic knowledge retrieval through entity-centric RAG to address rare entities and innovatively proposes an entity-title collaborative corpus, and (2) employs contrastive correction tasks to mitigate LLM over-correction tendencies. Furthermore, we propose MDCQC, a Multi-Domain Chinese Query Correction benchmark to test the model’s entity correction capabilities. Extensive experiments on several datasets show that RACQC significantly outperforms existing baselines in CSC tasks. Specifically, RACQC achieves a maximum improvement of +9.92% on the search scenario benchmark and +3.2% on the general-domain dataset under the F1 metric.
SAM Decoding: Speculative Decoding via Suffix Automaton
Yuxuan Hu | Ke Wang | Xiaokang Zhang | Fanjin Zhang | Cuiping Li | Hong Chen | Jing Zhang
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Yuxuan Hu | Ke Wang | Xiaokang Zhang | Fanjin Zhang | Cuiping Li | Hong Chen | Jing Zhang
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Speculative decoding (SD) has been demonstrated as an effective technique for lossless LLM inference acceleration. Retrieval-based SD methods, one kind of model-free method, have yielded promising speedup, but they often rely on single retrieval resources, inefficient retrieval methods, and are constrained to certain tasks. This paper presents a novel retrieval-based speculative decoding method that adapts the suffix automaton (SAM) for efficient and accurate draft generation by utilizing the generating text sequence and static text corpus. Unlike existing n-gram matching methods, SAM-Decoding finds the exact longest suffix match, achieving an average time complexity of O(1) per generation step of SAM update and suffix retrieval.It can also integrate with existing methods, adaptively selecting a draft generation strategy based on match length to generalize to broader domains. Extensive experiments on Spec-Bench show that our method is 18% faster than other retrieval-based SD methods. Additionally, when combined with advanced EAGLE-2, it provides an additional speedup of 3.28% – 11.13% across various-sized LLM backbones.