LPNL: Scalable Link Prediction with Large Language Models

Baolong Bi, Shenghua Liu, Yiwei Wang, Lingrui Mei, Xueqi Cheng


Abstract
Exploring the application of large language models (LLMs) to graph learning is an emerging endeavor. However, the vast amount of information inherent in large graphs poses significant challenges to graph learning with LLMs. This work focuses on the link prediction task and introduces **LPNL** (Link Prediction via Natural Language), a framework based on large language models designed for scalable link prediction on large-scale heterogeneous graphs. We design novel prompts for link prediction that articulate graph details in natural language. We propose a two-stage sampling pipeline to extract crucial information from the graphs, and a divide-and-conquer strategy to control the input tokens within predefined limits, addressing the challenge of overwhelming information. We fine-tune a T5 model based on our self-supervised learning designed for link prediction. Extensive experimental results demonstrate that LPNL outperforms multiple advanced baselines in link prediction tasks on large-scale graphs.
Anthology ID:
2024.findings-acl.215
Volume:
Findings of the Association for Computational Linguistics ACL 2024
Month:
August
Year:
2024
Address:
Bangkok, Thailand and virtual meeting
Editors:
Lun-Wei Ku, Andre Martins, Vivek Srikumar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
3615–3625
Language:
URL:
https://aclanthology.org/2024.findings-acl.215
DOI:
Bibkey:
Cite (ACL):
Baolong Bi, Shenghua Liu, Yiwei Wang, Lingrui Mei, and Xueqi Cheng. 2024. LPNL: Scalable Link Prediction with Large Language Models. In Findings of the Association for Computational Linguistics ACL 2024, pages 3615–3625, Bangkok, Thailand and virtual meeting. Association for Computational Linguistics.
Cite (Informal):
LPNL: Scalable Link Prediction with Large Language Models (Bi et al., Findings 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.findings-acl.215.pdf