Quyen Tran
2024
Lifelong Event Detection via Optimal Transport
Viet Dao
|
Van-Cuong Pham
|
Quyen Tran
|
Thanh-Thien Le
|
Linh Van Ngo
|
Thien Huu Nguyen
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Continual Event Detection (CED) poses a formidable challenge due to the catastrophic forgetting phenomenon, where learning new tasks (with new coming event types) hampers performance on previous ones. In this paper, we introduce a novel approach, Lifelong Event Detection via Optimal Transport (**LEDOT**), that leverages optimal transport principles to align the optimization of our classification module with the intrinsic nature of each class, as defined by their pre-trained language modeling. Our method integrates replay sets, prototype latent representations, and an innovative Optimal Transport component. Extensive experiments on MAVEN and ACE datasets demonstrate LEDOT’s superior performance, consistently outperforming state-of-the-art baselines. The results underscore LEDOT as a pioneering solution in continual event detection, offering a more effective and nuanced approach to addressing catastrophic forgetting in evolving environments.
Preserving Generalization of Language models in Few-shot Continual Relation Extraction
Quyen Tran
|
Nguyen Xuan Thanh
|
Nguyen Hoang Anh
|
Nam Le Hai
|
Trung Le
|
Linh Van Ngo
|
Thien Huu Nguyen
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Few-shot Continual Relations Extraction (FCRE) is an emerging and dynamic area of study where models can sequentially integrate knowledge from new relations with limited labeled data while circumventing catastrophic forgetting and preserving prior knowledge from pre-trained backbones. In this work, we introduce a novel method that leverages often-discarded language model heads. By employing these components via a mutual information maximization strategy, our approach helps maintain prior knowledge from the pre-trained backbone and strategically aligns the primary classification head, thereby enhancing model performance. Furthermore, we explore the potential of Large Language Models (LLMs), renowned for their wealth of knowledge, in addressing FCRE challenges. Our comprehensive experimental results underscore the efficacy of the proposed method and offer valuable insights for future work.
Search
Co-authors
- Linh Van Ngo 2
- Thien Huu Nguyen 2
- Viet Dao 1
- Van-Cuong Pham 1
- Thanh-Thien Le 1
- show all...