An Efficient Retrieval-Based Method for Tabular Prediction with LLM

Jie Wu, Mengshu Hou


Abstract
Tabular prediction, a well-established problem in machine learning, has consistently garnered significant research attention within academia and industry. Recently, with the rapid development of large language models (LLMs), there has been increasing exploration of how to apply LLMs to tabular prediction tasks. Many existing methods, however, typically rely on extensive pre-training or fine-tuning of LLMs, which demands considerable computational resources. To avoid this, we propose a retrieval-based approach that utilizes the powerful capabilities of LLMs in representation, comprehension, and inference. Our approach eliminates the need for training any modules or performing data augmentation, depending solely on information from target dataset. Experimental results reveal that, even without specialized training for tabular data, our method exhibits strong predictive performance on tabular prediction task, affirming its practicality and effectiveness.
Anthology ID:
2025.coling-main.663
Volume:
Proceedings of the 31st International Conference on Computational Linguistics
Month:
January
Year:
2025
Address:
Abu Dhabi, UAE
Editors:
Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Venue:
COLING
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
9917–9925
Language:
URL:
https://aclanthology.org/2025.coling-main.663/
DOI:
Bibkey:
Cite (ACL):
Jie Wu and Mengshu Hou. 2025. An Efficient Retrieval-Based Method for Tabular Prediction with LLM. In Proceedings of the 31st International Conference on Computational Linguistics, pages 9917–9925, Abu Dhabi, UAE. Association for Computational Linguistics.
Cite (Informal):
An Efficient Retrieval-Based Method for Tabular Prediction with LLM (Wu & Hou, COLING 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.coling-main.663.pdf