Xiu Yan
2025
Patchwise Cooperative Game-based Interpretability Method for Large Vision-language Models
Yao Zhu | Yunjian Zhang | Zizhe Wang | Xiu Yan | Peng Sun | Xiangyang Ji
Transactions of the Association for Computational Linguistics, Volume 13
Yao Zhu | Yunjian Zhang | Zizhe Wang | Xiu Yan | Peng Sun | Xiangyang Ji
Transactions of the Association for Computational Linguistics, Volume 13
Amidst the rapid advancement of artificial intelligence, research on large vision-language models (LVLMs) has emerged as a pivotal area. However, understanding their internal mechanisms remains challenging due to the limitations of existing interpretability methods, especially regarding faithfulness and plausibility. To address this, we first construct a human response interpretability dataset that evaluates the plausibility of model explanations by comparing the attention regions between the model and humans when answering the same questions. We then propose a patchwise cooperative game-based interpretability method for LVLMs, which employs Shapley values to quantify the impact of individual image patches on generation likelihood and enhances computational efficiency through a single input approximation approach. Experimental results demonstrate our method’s faithfulness, plausibility, and robustness. Our method provides researchers with deeper insights into model behavior, allowing for an examination of the specific image regions each layer relies on during response generation, ultimately enhancing model reliability. Our code is available at https://github.com/ZY123-GOOD/Patchwise_Cooperative.
2024
Unleashing the Potential of Large Language Models through Spectral Modulation
Peng Sun | Yao Zhu | Yunjian Zhang | Xiu Yan | Zizhe Wang | Xiangyang Ji
Findings of the Association for Computational Linguistics: EMNLP 2024
Peng Sun | Yao Zhu | Yunjian Zhang | Xiu Yan | Zizhe Wang | Xiangyang Ji
Findings of the Association for Computational Linguistics: EMNLP 2024
Large Language Models (LLMs) have demonstrated impressive capabilities across various domains, garnering significant attention from both academia and industry. However, enhancing the performance of LLMs typically requires scaling up model sizes or fine-tuning with additional datasets, which results in substantial computational costs. This paper poses an intriguing question: Can we improve the performance of LLMs without additional training? Drawing inspiration from signal processing principles, which suggest that noise often resides in high-frequency components while low-frequency components carry the essence of signals, we propose uncovering untapped potential in LLMs from a frequency perspective. We hypothesize that the high-frequency components in the weight matrices of LLMs’ linear layers may conceal noise that interferes with predictive accuracy. Therefore, we propose conducting spectral modulation in the parameter space of LLMs, which can seamlessly integrate with various models in a plug-and-play manner. Extensive experiments have demonstrated the superiority of our approach, with spectral modulation yielding an average performance improvement of up to 10.12%.