Yizhuo Zhang
2024
Can LLM Graph Reasoning Generalize beyond Pattern Memorization?
Yizhuo Zhang
|
Heng Wang
|
Shangbin Feng
|
Zhaoxuan Tan
|
Xiaochuang Han
|
Tianxing He
|
Yulia Tsvetkov
Findings of the Association for Computational Linguistics: EMNLP 2024
Large language models (LLMs) demonstrate great potential for problems with implicit graphical structures, while recent works seek to enhance the graph reasoning capabilities of LLMs through specialized instruction tuning. The resulting “graph LLMs” are evaluated with in-distribution settings only, thus it remains underexplored whether LLMs are learning generalizable graph reasoning skills or merely memorizing patterns in the synthetic training data. To this end, we propose the NLGift benchmark, an evaluation suite of LLM graph reasoning generalization: whether LLMs could go beyond semantic, numeric, structural, reasoning patterns in the synthetic training data and improve utility on real-world graph-based tasks. Extensive experiments with two LLMs across four graph reasoning tasks demonstrate that while generalization on simple patterns (semantic, numeric) is somewhat satisfactory, LLMs struggle to generalize across reasoning and real-world patterns, casting doubt on the benefit of synthetic graph tuning for real-world tasks with underlying network structures. We explore three strategies to improve LLM graph reasoning generalization, and we find that while post-training alignment is most promising for real-world tasks, empowering LLM graph reasoning to go beyond pattern memorization remains an open research question.
Search
Co-authors
- Heng Wang 1
- Shangbin Feng 1
- Zhaoxuan Tan 1
- Xiaochuang Han 1
- Tianxing He 1
- show all...