Anton Bulle Labate


2026

This paper investigates whether injecting semantic structural knowledge of low-resource or unfamiliar languages into Large Language Models (LLMs) enhances performance on downstream Text-to-SQL tasks. We evaluate our approach on Galician, a Romance low-resource language, and, to demonstrate its generality, also on Guarani, a (very) low-resource language of an entirely distinct linguistic profile. Our empirical results show that semantically-aware models consistently outperform baselines across all benchmark metrics.