Severity Classification and Dysarthric Speech Detection using Self-Supervised Representations

Sanjay B, Priyadharshini M.k, Vijayalakshmi P, Nagarajan T


Abstract
Automatic detection and classification of dysarthria severity from speech provides a non-invasive and efficient diagnostic tool, offering clinicians valuable insights to guide treatment and therapy decisions. Our study evaluated two pre-trained models—wav2vec2-BASE and distilALHuBERT, for feature extraction to build speech detection and severity-level classification systems for dysarthric speech. We conducted experiments on the TDSC dataset using two approaches: a machine learning model (support vector machine, SVM) and a deep learning model (convolutional neural network, CNN). Our findings showed that features derived from distilALHuBERT significantly outperformed those from wav2vec2-BASE in both dysarthric speech detection and severity classification tasks. Notably, the distilALHuBERT features achieved 99% accuracy in automatic detection and 95% accuracy in severity classification, surpassing the performance of wav2vec2 features.
Anthology ID:
2024.icon-1.74
Volume:
Proceedings of the 21st International Conference on Natural Language Processing (ICON)
Month:
December
Year:
2024
Address:
AU-KBC Research Centre, Chennai, India
Editors:
Sobha Lalitha Devi, Karunesh Arora
Venue:
ICON
SIG:
Publisher:
NLP Association of India (NLPAI)
Note:
Pages:
621–628
Language:
URL:
https://aclanthology.org/2024.icon-1.74/
DOI:
Bibkey:
Cite (ACL):
Sanjay B, Priyadharshini M.k, Vijayalakshmi P, and Nagarajan T. 2024. Severity Classification and Dysarthric Speech Detection using Self-Supervised Representations. In Proceedings of the 21st International Conference on Natural Language Processing (ICON), pages 621–628, AU-KBC Research Centre, Chennai, India. NLP Association of India (NLPAI).
Cite (Informal):
Severity Classification and Dysarthric Speech Detection using Self-Supervised Representations (B et al., ICON 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.icon-1.74.pdf