AI-Tutor: Interactive Learning of Ancient Knowledge from Low-Resource Languages

Siddhartha Dalal, Rahul Aditya, Vethavikashini Chithrra Raghuram, Prahlad Koratamaddi


Abstract
Many low-resource languages, such as Prakrit, present significant linguistic complexities and have limited modern-day resources. These languages often have multiple derivatives; for example, Prakrit, a language in use by masses around 2500 years ago for 500 years, includes Pali and Gandhari, which encompass a vast body of Buddhist literature, as well as Ardhamagadhi, rich in Jain literature. Despite these challenges, these languages are invaluable for their historical, religious, and cultural insights needed by non-language experts and others.To explore and understand the deep knowledge within these ancient texts for non-language experts, we propose a novel approach: translating multiple dialects of the parent language into a contemporary language and then enabling them to interact with the system in their native language, including English, Hindi, French and German, through a question-and-answer interface built on Large Language Models. We demonstrate the effectiveness of this novel AI-Tutor system by focusing on Ardhamagadhi and Pali.
Anthology ID:
2024.wat-1.5
Volume:
Proceedings of the Eleventh Workshop on Asian Translation (WAT 2024)
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Editors:
Toshiaki Nakazawa, Isao Goto
Venue:
WAT
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
56–66
Language:
URL:
https://aclanthology.org/2024.wat-1.5
DOI:
Bibkey:
Cite (ACL):
Siddhartha Dalal, Rahul Aditya, Vethavikashini Chithrra Raghuram, and Prahlad Koratamaddi. 2024. AI-Tutor: Interactive Learning of Ancient Knowledge from Low-Resource Languages. In Proceedings of the Eleventh Workshop on Asian Translation (WAT 2024), pages 56–66, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
AI-Tutor: Interactive Learning of Ancient Knowledge from Low-Resource Languages (Dalal et al., WAT 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.wat-1.5.pdf
Supplementary material:
 2024.wat-1.5.SupplementaryMaterial.txt
Supplementary material:
 2024.wat-1.5.SupplementaryMaterial.zip