Mícheál J. Ó Meachair
2025
Gaeilge Bhriste ó Shamhlacha Cliste: How Clever Are LLMs When Translating Irish Text?
Teresa Clifford
|
Abigail Walsh
|
Brian Davis
|
Mícheál J. Ó Meachair
Proceedings of the 5th Celtic Language Technology Workshop
Large Language Models have been widely adopted in NLP tasks and applications, how- ever, their ability to accurately process Irish and other minority languages has not been fully explored. In this paper we describe prelim- inary experiments examining the capacity of publicly-available machine translation engines (Google Translate, Microsoft Bing, and eTrans- lation) and prompt-based AI systems systems (ChatGPT 3.5, Llama 2) for translating and handling challenging language features of Irish. A hand-crafted selection of challenging Irish language features were incorporated into trans- lation prompts, and the output from each model was examined by a human evaluator. The re- sults of these experiments indicate that these LLM-based models still struggle with translat- ing rare linguistic phenomena and ambiguous constructions. This preliminary analysis helps to inform further research in this field, pro- viding a simple ranking of publicly-available models, and indicating which language features require particular attention when evaluating model capacity.
2022
gaBERT — an Irish Language Model
James Barry
|
Joachim Wagner
|
Lauren Cassidy
|
Alan Cowap
|
Teresa Lynn
|
Abigail Walsh
|
Mícheál J. Ó Meachair
|
Jennifer Foster
Proceedings of the Thirteenth Language Resources and Evaluation Conference
The BERT family of neural language models have become highly popular due to their ability to provide sequences of text with rich context-sensitive token encodings which are able to generalise well to many NLP tasks. We introduce gaBERT, a monolingual BERT model for the Irish language. We compare our gaBERT model to multilingual BERT and the monolingual Irish WikiBERT, and we show that gaBERT provides better representations for a downstream parsing task. We also show how different filtering criteria, vocabulary size and the choice of subword tokenisation model affect downstream performance. We compare the results of fine-tuning a gaBERT model with an mBERT model for the task of identifying verbal multiword expressions, and show that the fine-tuned gaBERT model also performs better at this task. We release gaBERT and related code to the community.
Search
Fix data
Co-authors
- Abigail Walsh 2
- James Barry 1
- Lauren Cassidy 1
- Teresa Clifford 1
- Alan Cowap 1
- show all...