Jin-Tao Tang
2025
R2A-TLS: Reflective Retrieval-Augmented Timeline Summarization with Causal-Semantic Integration
Chenlong Bao | Shijie Li | Minghao Hu | Ming Qiao | Bin Zhang | Jin-Tao Tang | Shasha Li | Ting Wang
Findings of the Association for Computational Linguistics: EMNLP 2025
Chenlong Bao | Shijie Li | Minghao Hu | Ming Qiao | Bin Zhang | Jin-Tao Tang | Shasha Li | Ting Wang
Findings of the Association for Computational Linguistics: EMNLP 2025
Open-domain timeline summarization (TLS) faces challenges from information overload and data sparsity when processing large-scale textual streams. Existing methods struggle to capture coherent event narratives due to fragmented descriptions and often accumulate noise through iterative retrieval strategies that lack effective relevance evaluation. This paper proposes: Reflective Retrieval-Augmented Timeline Summarization with Causal-Semantic Intergration, which offers a novel perspective for open-domain TLS by time point completion and event element completion. R2A-TLS establishes an initial retrieval, reflection, and deep retrieval system that reduces noise through a double filtering mechanism that iteratively generates a timeline for each text which passes the filtering. Then, the system reflects on the initial timeline with the aim of identifying information gaps through causal chain analysis and FrameNet based element validation. These gaps are reformulated into targeted queries to trigger deep retrieval for refining timeline coherence and density. Empirical evaluation on Open-TLS dataset reveals that our approach outperforms the best prior published approaches.
基于大模型增强的两阶段高效事件共指消解方法
Wu Yaozong | Shuai Qi | Fangyuan Wang | Chenlong Bao | Jin-Tao Tang
Proceedings of the 24th China National Conference on Computational Linguistics (CCL 2025)
Wu Yaozong | Shuai Qi | Fangyuan Wang | Chenlong Bao | Jin-Tao Tang
Proceedings of the 24th China National Conference on Computational Linguistics (CCL 2025)
"本文针对两阶段事件共指消解方法存在的触发词词目启发机制缺乏同义词聚类能力和小模型理解触发词指代事件能力有限等问题,提出了一种基于大模型增强的两阶段高效的事件共指消解方法,一阶段引入大模型进行同义词聚类,二阶段大模型提供触发词解释文本增强小模型。此外,设计了引导小模型侧重触发词特征向量的损失函数。本文方法在保持近似线性时间复杂度的同时,在ECB+和GVC数据集上的CoNLLF1得分分别提升了2.9和8.0。"