Multi-accent Speech Separation with One Shot Learning

Kuan Po Huang, Yuan-Kuei Wu, Hung-yi Lee


Abstract
Speech separation is a problem in the field of speech processing that has been studied in full swing recently. However, there has not been much work studying a multi-accent speech separation scenario. Unseen speakers with new accents and noise aroused the domain mismatch problem which cannot be easily solved by conventional joint training methods. Thus, we applied MAML and FOMAML to tackle this problem and obtained higher average Si-SNRi values than joint training on almost all the unseen accents. This proved that these two methods do have the ability to generate well-trained parameters for adapting to speech mixtures of new speakers and accents. Furthermore, we found out that FOMAML obtains similar performance compared to MAML while saving a lot of time.
Anthology ID:
2021.metanlp-1.7
Volume:
Proceedings of the 1st Workshop on Meta Learning and Its Applications to Natural Language Processing
Month:
August
Year:
2021
Address:
Online
Editors:
Hung-Yi Lee, Mitra Mohtarami, Shang-Wen Li, Di Jin, Mandy Korpusik, Shuyan Dong, Ngoc Thang Vu, Dilek Hakkani-Tur
Venue:
MetaNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
59–66
Language:
URL:
https://aclanthology.org/2021.metanlp-1.7
DOI:
10.18653/v1/2021.metanlp-1.7
Bibkey:
Cite (ACL):
Kuan Po Huang, Yuan-Kuei Wu, and Hung-yi Lee. 2021. Multi-accent Speech Separation with One Shot Learning. In Proceedings of the 1st Workshop on Meta Learning and Its Applications to Natural Language Processing, pages 59–66, Online. Association for Computational Linguistics.
Cite (Informal):
Multi-accent Speech Separation with One Shot Learning (Huang et al., MetaNLP 2021)
Copy Citation:
PDF:
https://aclanthology.org/2021.metanlp-1.7.pdf