Benchmarking Language-agnostic Intent Classification for Virtual Assistant Platforms

Gengyu Wang, Cheng Qian, Lin Pan, Haode Qi, Ladislav Kunc, Saloni Potdar


Abstract
Current virtual assistant (VA) platforms are beholden to the limited number of languages they support. Every component, such as the tokenizer and intent classifier, is engineered for specific languages in these intricate platforms. Thus, supporting a new language in such platforms is a resource-intensive operation requiring expensive re-training and re-designing. In this paper, we propose a benchmark for evaluating language-agnostic intent classification, the most critical component of VA platforms. To ensure the benchmarking is challenging and comprehensive, we include 29 public and internal datasets across 10 low-resource languages and evaluate various training and testing settings with consideration of both accuracy and training time. The benchmarking result shows that Watson Assistant, among 7 commercial VA platforms and pre-trained multilingual language models (LMs), demonstrates close-to-best accuracy with the best accuracy-training time trade-off.
Anthology ID:
2022.mia-1.7
Volume:
Proceedings of the Workshop on Multilingual Information Access (MIA)
Month:
July
Year:
2022
Address:
Seattle, USA
Editors:
Akari Asai, Eunsol Choi, Jonathan H. Clark, Junjie Hu, Chia-Hsuan Lee, Jungo Kasai, Shayne Longpre, Ikuya Yamada, Rui Zhang
Venue:
MIA
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
69–76
Language:
URL:
https://aclanthology.org/2022.mia-1.7
DOI:
10.18653/v1/2022.mia-1.7
Bibkey:
Cite (ACL):
Gengyu Wang, Cheng Qian, Lin Pan, Haode Qi, Ladislav Kunc, and Saloni Potdar. 2022. Benchmarking Language-agnostic Intent Classification for Virtual Assistant Platforms. In Proceedings of the Workshop on Multilingual Information Access (MIA), pages 69–76, Seattle, USA. Association for Computational Linguistics.
Cite (Informal):
Benchmarking Language-agnostic Intent Classification for Virtual Assistant Platforms (Wang et al., MIA 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.mia-1.7.pdf
Code
 posuer/benchmark-multilingual-intent-classification