Zhihui Yang
2025
CCL25-Eval任务四系统报告:基于多策略知识融合的叙实性推理方法研究
Hongyu Li | Zhihui Yang | Renfen Hu
Proceedings of the 24th China National Conference on Computational Linguistics (CCL 2025)
Hongyu Li | Zhihui Yang | Renfen Hu
Proceedings of the 24th China National Conference on Computational Linguistics (CCL 2025)
"FIE2025任务旨在使用大语言模型对文本及相关假设进行叙实性推理。我们参加了微调和非微调两个赛道,分别在人工数据集和自然数据集上采用提示词优化和词表RAG策略融合语言学知识,并利用模型集成投票方法提升判断准确率。评测结果显示,我们的方法在非微调赛道取得了0.9351的成绩,在微调赛道取得了0.9261的成绩,均位列第三名。"
Does Visual Grounding Enhance the Understanding of Embodied Knowledge in Large Language Models?
Zhihui Yang | Yupei Wang | Kaijie Mo | Zhe Zhao | Renfen Hu
Findings of the Association for Computational Linguistics: EMNLP 2025
Zhihui Yang | Yupei Wang | Kaijie Mo | Zhe Zhao | Renfen Hu
Findings of the Association for Computational Linguistics: EMNLP 2025
Despite significant progress in multimodal language models (LMs), it remains unclear whether visual grounding enhances their understanding of embodied knowledge compared to text-only models. To address this question, we propose a novel embodied knowledge understanding benchmark based on the perceptual theory from psychology, encompassing visual, auditory, tactile, gustatory, olfactory external senses, and interoception. The benchmark assesses the models’ perceptual abilities across different sensory modalities through vector comparison and question-answering tasks with over 1,700 questions. By comparing 30 state-of-the-art LMs, we surprisingly find that vision-language models (VLMs) do not outperform text-only models in either task. Moreover, the models perform significantly worse in the visual dimension compared to other sensory dimensions. Further analysis reveals that the vector representations are easily influenced by word form and frequency, and the models struggle to answer questions involving spatial perception and reasoning. Our findings underscore the need for more effective integration of embodied knowledge in LMs to enhance their understanding of the physical world.