Yu-Sheng Huang


2025

pdf bib
Applying Whisper Fine-tuning and Branchformer to Hakka Speech Recognition
Yu-Sheng Huang | Wei-Cheng Hong | Xin-Yu Chen | Szu-Yin Lin
Proceedings of the 37th Conference on Computational Linguistics and Speech Processing (ROCLING 2025)

This study addresses the FSR 2025 Hakka speech recognition task by comparing two strategies: fine-tuning large pre-trained models and training from scratch. For character (Hanzi) recognition, we fine-tuned five different scales of the Whisper model, with large-v3-turbo achieving a 7.55% CER on the test set. For Pinyin recognition, a Branchformer model was compared against a LoRA fine-tuned Whisper-small, yielding WERs of 4.7% and 6.5% on the test set, respectively. Speed perturbation was the primary method used for data augmentation in our pre-processing pipeline.

2019

pdf bib
A Hybrid Approach of Deep Semantic Matching and Deep Rank for Context Aware Question Answer System
Shu-Yi Xie | Chia-Hao Chang | Zhi Zhang | Yang Mo | Lian-Xin Jiang | Yu-Sheng Huang | Jian-Ping Shen
Proceedings of the 31st Conference on Computational Linguistics and Speech Processing (ROCLING 2019)

pdf bib
A Real-World Human-Machine Interaction Platform in Insurance Industry
Wei Tan | Chia-Hao Chang | Yang Mo | Lian-Xin Jiang | Gen Li | Xiao-Long Hou | Chu Chen | Yu-Sheng Huang | Meng-Yuan Huang | Jian-Ping Shen
Proceedings of the 31st Conference on Computational Linguistics and Speech Processing (ROCLING 2019)