Jiahao Teng
2025
System Report for CCL25-Eval Task 8: Structured ICD Coding with LLM-Augmented Learning and Group-specific Classifiers
Bo Wang | Kaiyuan Zhang | Chong Feng | Ge Shi | Jinhua Ye | Jiahao Teng | Shouzhen Wang | Fanqing Meng | Changsen Yuan | Yan Zhuang
Proceedings of the 24th China National Conference on Computational Linguistics (CCL 2025)
Bo Wang | Kaiyuan Zhang | Chong Feng | Ge Shi | Jinhua Ye | Jiahao Teng | Shouzhen Wang | Fanqing Meng | Changsen Yuan | Yan Zhuang
Proceedings of the 24th China National Conference on Computational Linguistics (CCL 2025)
"The International Classification of Diseases (ICD) provides a standardized framework for encoding diagnoses, serving critical roles in clinical scenarios. Automatic ICD coding aims to assign formalized diagnostic codes to medical records for documentation and analysis, which is challenged by an extremely large and imbalanced label space, noisy and heterogeneous clinical text,and the need for interpretability. In this paper, we propose a structured multi-class classification framework that partitions diseases into clinically coherent groups, enabling group-specific dataaugmentation and supervision. Our method combines input compression with generative and discriminative fine-tuning strategies tailored to primary and secondary diagnoses, respectively.On the CCL2025-Eval Task 8 benchmark for Chinese electronic medical records, our approach ranked first in the final evaluation."
Mitigating the Discrepancy Between Video and Text Temporal Sequences: A Time-Perception Enhanced Video Grounding method for LLM
Xuefen Li | Bo Wang | Ge Shi | Chong Feng | Jiahao Teng
Proceedings of the 31st International Conference on Computational Linguistics
Xuefen Li | Bo Wang | Ge Shi | Chong Feng | Jiahao Teng
Proceedings of the 31st International Conference on Computational Linguistics
Existing video LLMs typically excel at capturing the overall description of a video but lack the ability to demonstrate an understanding of temporal dynamics and a fine-grained grasp of localized content within the video. In this paper, we propose a Time-Perception Enhanced Video Grounding via Boundary Perception and Temporal Reasoning aimed at mitigating LLMs’ difficulties in understanding the discrepancies between video and text temporality. Specifically, to address the inherent biases in current datasets, we design a series of boundary-perception tasks to enable LLMs to capture accurate video temporality. To tackle LLMs’ insufficient understanding of temporal information, we develop specialized tasks for boundary perception and temporal relationship reasoning to deepen LLMs’ perception of video temporality. Our experimental results show significant improvements across three datasets: ActivityNet, Charades, and DiDeMo (achieving up to 11.2% improvement on R@0.3), demonstrating the effectiveness of our proposed temporal awareness-enhanced data construction method.