2025
pdf
bib
abs
STAND-Guard: A Small Task-Adaptive Content Moderation Model
Minjia Wang
|
Pingping Lin
|
Siqi Cai
|
Shengnan An
|
Shengjie Ma
|
Zeqi Lin
|
Congrui Huang
|
Bixiong Xu
Proceedings of the 31st International Conference on Computational Linguistics: Industry Track
Content moderation, the process of reviewing and monitoring the safety of generated content, is important for development of welcoming online platforms and responsible large language models. Content moderation contains various tasks, each with its unique requirements tailored to specific scenarios. Therefore, it is crucial to develop a model that can be easily adapted to novel or customized content moderation tasks accurately without extensive model tuning. This paper presents STAND-Guard, a Small Task-Adaptive coNtent moDeration model. The basic motivation is: by performing instruct tuning on various content moderation tasks, we can unleash the power of small language models (SLMs) on unseen (out-of-distribution) content moderation tasks. We also carefully study the effects of training tasks and model size on the efficacy of cross-task fine-tuning mechanism. Experiments demonstrate STAND-Guard is comparable to GPT-3.5-Turbo across over 40 public datasets, as well as proprietary datasets derived from real-world business scenarios. Remarkably, STAND-Guard achieved nearly equivalent results to GPT-4-Turbo on unseen English binary classification tasks.
2024
pdf
bib
abs
Can LLMs Learn From Mistakes? An Empirical Study on Reasoning Tasks
Shengnan An
|
Zexiong Ma
|
Siqi Cai
|
Zeqi Lin
|
Nanning Zheng
|
Jian-Guang Lou
|
Weizhu Chen
Findings of the Association for Computational Linguistics: EMNLP 2024
Towards enhancing the chain-of-thought (CoT) reasoning of large language models (LLMs), much existing work has revealed the effectiveness of straightforward learning on annotated/generated CoT paths. However, there is less evidence yet that reasoning capabilities can be enhanced through a reverse learning process, i.e., learning from potential mistakes in reasoning. To investigate whether LLMs can learn from mistakes, we construct mistake-correction datasets, using GPT-4 to identify and correct the mistakes in inaccurate CoTs. With these mistake-correction datasets, we fine-tune open-source LLMs and arrive at the following conclusions. (1) LLMs can indeed learn from mistakes to enhance their CoT reasoning performances. (2) Compared to CoT data, the mistake-correction data provides additional knowledge on the explanations and reasons for the potential mistakes in CoTs, which consistently contributes to the effectiveness of learning from mistakes. (3) Evolution techniques, especially the correction-centric evolution we introduced, can further enhance the effectiveness of learning from mistakes.
2021
pdf
bib
Proceedings of the 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue
Haizhou Li
|
Gina-Anne Levow
|
Zhou Yu
|
Chitralekha Gupta
|
Berrak Sisman
|
Siqi Cai
|
David Vandyke
|
Nina Dethlefs
|
Yan Wu
|
Junyi Jessy Li
Proceedings of the 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue