Chinese Grammatical Error Correction via Large Language Model Guided Optimization Training

Liu Xiao, Li Ying, Yu Zhengtao


Abstract
“Pre-trained language model-based methods for Chinese Grammatical Error Correction (CGEC)are categorized into Seq2Seq and Seq2Edit types. However, both Seq2Seq and Seq2Edit mod-els depend on high-quality training data significantly. Considering the strong generation andinference ability of large language models (LLMs), we propose a large language model-guidedoptimization training method to exploit LLMs to extract error knowledge to optimize the tradi-tional CGEC model training process. On the one hand, we use error types and confusion sets asextra knowledge to guide LLMs to generate diverse pseudo data, thus extending the error distri-bution of our training data. On the other hand, LLMs are utilized to infer the predicted resultsfrom our CGEC models and obtain the re-training data, thus iteratively optimizing our pre-trainedCGEC models. Experiments on two benchmark datasets show that our LLMs-guided optimiza-tion method with small-scale training data can achieve comparable results with baseline modelswith large-scale training data. Detailed comparison experiments demonstrate that both the earlydeviser pseudo data and the later re-training data are extremely useful for traditional CGEC modeloptimization training, and can benefit from each other. We will release our code and prompts athttps://github.com/SakuraAcedia/llm-cgec-got to facilitate future work.”
Anthology ID:
2024.ccl-1.105
Volume:
Proceedings of the 23rd Chinese National Conference on Computational Linguistics (Volume 1: Main Conference)
Month:
July
Year:
2024
Address:
Taiyuan, China
Editors:
Maosong Sun, Jiye Liang, Xianpei Han, Zhiyuan Liu, Yulan He
Venue:
CCL
SIG:
Publisher:
Chinese Information Processing Society of China
Note:
Pages:
1366–1380
Language:
English
URL:
https://aclanthology.org/2024.ccl-1.105/
DOI:
Bibkey:
Cite (ACL):
Liu Xiao, Li Ying, and Yu Zhengtao. 2024. Chinese Grammatical Error Correction via Large Language Model Guided Optimization Training. In Proceedings of the 23rd Chinese National Conference on Computational Linguistics (Volume 1: Main Conference), pages 1366–1380, Taiyuan, China. Chinese Information Processing Society of China.
Cite (Informal):
Chinese Grammatical Error Correction via Large Language Model Guided Optimization Training (Xiao et al., CCL 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.ccl-1.105.pdf