Kevin Miller
2026
Hearing Between the Lines: Unlocking the Reasoning Power of LLMs for Speech Evaluation
Arjun Chandra | Kevin Miller | Venkatesh Ravichandran | Constantinos Papayiannis | Venkatesh Saligrama
Findings of the Association for Computational Linguistics: EACL 2026
Arjun Chandra | Kevin Miller | Venkatesh Ravichandran | Constantinos Papayiannis | Venkatesh Saligrama
Findings of the Association for Computational Linguistics: EACL 2026
Large Language Model (LLM) judges exhibit strong reasoning capabilities but are limited to textual content. This leaves current automatic Speech-to-Speech (S2S) evaluation methods reliant on opaque and expensive Audio Language Models (ALMs). In this work, we propose TRACE (Textual Reasoning over Audio Cues for Evaluation), a novel framework that enables LLM judges to reason over audio cues to achieve cost-efficient and human-aligned S2S evaluation. To demonstrate the strength of the framework, we first introduce a Human Chain-of-Thought (HCoT) annotation protocol to improve the diagnostic capability of existing judge benchmarks by separating evaluation into explicit dimensions: content (C), voice quality (VQ), and paralinguistics (P). Using this data, TRACE constructs a textual blueprint of inexpensive audio signals and prompts an LLM to render dimension-wise judgments, fusing them into an overall rating via a deterministic policy. TRACE achieves higher agreement with human raters than ALMs and transcript-only LLM judges while being significantly more cost-effective. We will release the HCoT annotations and the TRACE framework to enable scalable and human-aligned S2S evaluation.
2025
Scaling Up Temporal Domain Generalization via Temporal Experts Averaging
Aoming Liu | Kevin Miller | Venkatesh Saligrama | Kate Saenko | Boqing Gong | Ser-Nam Lim | Bryan A. Plummer
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Aoming Liu | Kevin Miller | Venkatesh Saligrama | Kate Saenko | Boqing Gong | Ser-Nam Lim | Bryan A. Plummer
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Temporal Domain Generalization (TDG) aims to generalize across temporal distribution shifts, e.g., lexical change over time. Prior work often addresses this by predicting future model weights. However, full model prediction is prohibitively expensive for even reasonably sized models. Thus, recent methods only predict the classifier layer, limiting generalization by failing to adjust other model components. To address this, we propose Temporal Expert Averaging (TEA), a novel and scalable TDG framework that updates the entire model using weight averaging to maximize generalization potential while minimizing computational costs. Our theoretical analysis guides us to two steps that enhance generalization to future domains. First, we create expert models with functional diversity yet parameter similarity by fine-tuning a domain-agnostic base model on individual temporal domains while constraining weight changes. Second, we optimize the bias-variance tradeoff through adaptive averaging coefficients derived from modeling temporal weight trajectories in a principal component subspace. Expert’s contributions are based on their projected proximity to future domains. Extensive experiments across 7 TDG benchmarks, 5 models, and 2 TDG settings shows TEA outperforms prior TDG methods by up to 69% while being up to 60x more efficient.