Guirong Bai
2022
Incremental Intent Detection for Medical Domain with Contrast Replay Networks
Guirong Bai
|
Shizhu He
|
Kang Liu
|
Jun Zhao
Findings of the Association for Computational Linguistics: ACL 2022
Conventional approaches to medical intent detection require fixed pre-defined intent categories. However, due to the incessant emergence of new medical intents in the real world, such requirement is not practical. Considering that it is computationally expensive to store and re-train the whole data every time new data and intents come in, we propose to incrementally learn emerged intents while avoiding catastrophically forgetting old intents. We first formulate incremental learning for medical intent detection. Then, we employ a memory-based method to handle incremental learning. We further propose to enhance the method with contrast replay networks, which use multilevel distillation and contrast objective to address training data imbalance and medical rare words respectively. Experiments show that the proposed method outperforms the state-of-the-art model by 5.7% and 9.1% of accuracy on two benchmarks respectively.
2020
Pre-trained Language Model Based Active Learning for Sentence Matching
Guirong Bai
|
Shizhu He
|
Kang Liu
|
Jun Zhao
|
Zaiqing Nie
Proceedings of the 28th International Conference on Computational Linguistics
Active learning is able to significantly reduce the annotation cost for data-driven techniques. However, previous active learning approaches for natural language processing mainly depend on the entropy-based uncertainty criterion, and ignore the characteristics of natural language. In this paper, we propose a pre-trained language model based active learning approach for sentence matching. Differing from previous active learning, it can provide linguistic criteria from the pre-trained language model to measure instances and help select more effective instances for annotation. Experiments demonstrate our approach can achieve greater accuracy with fewer labeled training instances.