2024
pdf
bib
A Psycholinguistic Evaluation of Language Models’ Sensitivity to Argument Roles
Eun-Kyoung Rosa Lee
|
Sathvik Nair
|
Naomi Feldman
Findings of the Association for Computational Linguistics: EMNLP 2024
pdf
bib
abs
Self-supervised speech representations display some human-like cross-linguistic perceptual abilities
Joselyn Rodriguez
|
Kamala Sreepada
|
Ruolan Leslie Famularo
|
Sharon Goldwater
|
Naomi Feldman
Proceedings of the 28th Conference on Computational Natural Language Learning
State of the art models in automatic speech recognition have shown remarkable improvements due to modern self-supervised (SSL) transformer-based architectures such as wav2vec 2.0 (Baevski et al., 2020). However, how these models encode phonetic information is still not well understood. We explore whether SSL speech models display a linguistic property that characterizes human speech perception: language specificity. We show that while wav2vec 2.0 displays an overall language specificity effect when tested on Hindi vs. English, it does not resemble human speech perception when tested on finer-grained differences in Hindi speech contrasts.
2021
pdf
bib
abs
A phonetic model of non-native spoken word processing
Yevgen Matusevych
|
Herman Kamper
|
Thomas Schatz
|
Naomi Feldman
|
Sharon Goldwater
Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume
Non-native speakers show difficulties with spoken word processing. Many studies attribute these difficulties to imprecise phonological encoding of words in the lexical memory. We test an alternative hypothesis: that some of these difficulties can arise from the non-native speakers’ phonetic perception. We train a computational model of phonetic learning, which has no access to phonology, on either one or two languages. We first show that the model exhibits predictable behaviors on phone-level and word-level discrimination tasks. We then test the model on a spoken word processing task, showing that phonology may not be necessary to explain some of the word processing effects observed in non-native speakers. We run an additional analysis of the model’s lexical representation space, showing that the two training languages are not fully separated in that space, similarly to the languages of a bilingual human speaker.
2020
pdf
bib
Modeling the learning of the Person Case Constraint
Adam Liter
|
Naomi Feldman
Proceedings of the Society for Computation in Linguistics 2020
2017
pdf
bib
abs
Rational Distortions of Learners’ Linguistic Input
Naomi Feldman
Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL 2017)
Language acquisition can be modeled as a statistical inference problem: children use sentences and sounds in their input to infer linguistic structure. However, in many cases, children learn from data whose statistical structure is distorted relative to the language they are learning. Such distortions can arise either in the input itself, or as a result of children’s immature strategies for encoding their input. This work examines several cases in which the statistical structure of children’s input differs from the language being learned. Analyses show that these distortions of the input can be accounted for with a statistical learning framework by carefully considering the inference problems that learners solve during language acquisition
pdf
bib
abs
Evaluating Low-Level Speech Features Against Human Perceptual Data
Caitlin Richter
|
Naomi H. Feldman
|
Harini Salgado
|
Aren Jansen
Transactions of the Association for Computational Linguistics, Volume 5
We introduce a method for measuring the correspondence between low-level speech features and human perception, using a cognitive model of speech perception implemented directly on speech recordings. We evaluate two speaker normalization techniques using this method and find that in both cases, speech features that are normalized across speakers predict human data better than unnormalized speech features, consistent with previous research. Results further reveal differences across normalization methods in how well each predicts human data. This work provides a new framework for evaluating low-level representations of speech on their match to human perception, and lays the groundwork for creating more ecologically valid models of speech perception.
pdf
bib
abs
Learning an Input Filter for Argument Structure Acquisition
Laurel Perkins
|
Naomi Feldman
|
Jeffrey Lidz
Proceedings of the 7th Workshop on Cognitive Modeling and Computational Linguistics (CMCL 2017)
How do children learn a verb’s argument structure when their input contains nonbasic clauses that obscure verb transitivity? Here we present a new model that infers verb transitivity by learning to filter out non-basic clauses that were likely parsed in error. In simulations with child-directed speech, we show that this model accurately categorizes the majority of 50 frequent transitive, intransitive and alternating verbs, and jointly learns appropriate parameters for filtering parsing errors. Our model is thus able to filter out problematic data for verb learning without knowing in advance which data need to be filtered.
2016
pdf
bib
Joint Word Segmentation and Phonetic Category Induction
Micha Elsner
|
Stephanie Antetomaso
|
Naomi Feldman
Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
2015
pdf
bib
Why discourse affects speakers’ choice of referring expressions
Naho Orita
|
Eliana Vornov
|
Naomi Feldman
|
Hal Daumé III
Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)
2014
pdf
bib
Quantifying the role of discourse topicality in speakers’ choices of referring expressions
Naho Orita
|
Naomi Feldman
|
Jordan Boyd-Graber
|
Eliana Vornov
Proceedings of the Fifth Workshop on Cognitive Modeling and Computational Linguistics
pdf
bib
Weak semantic context helps phonetic learning in a model of infant language acquisition
Stella Frank
|
Naomi H. Feldman
|
Sharon Goldwater
Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
2013
pdf
bib
A Joint Learning Model of Word Segmentation, Lexical Acquisition, and Phonetic Variability
Micha Elsner
|
Sharon Goldwater
|
Naomi Feldman
|
Frank Wood
Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing