@inproceedings{oh-etal-2022-applicability,
title = "Applicability of Pretrained Language Models: Automatic Screening for Children{'}s Language Development Level",
author = "Oh, Byoung-doo and
Lee, Yoon-koung and
Kim, Yu-seop",
editor = "Biester, Laura and
Demszky, Dorottya and
Jin, Zhijing and
Sachan, Mrinmaya and
Tetreault, Joel and
Wilson, Steven and
Xiao, Lu and
Zhao, Jieyu",
booktitle = "Proceedings of the Second Workshop on NLP for Positive Impact (NLP4PI)",
month = dec,
year = "2022",
address = "Abu Dhabi, United Arab Emirates (Hybrid)",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.nlp4pi-1.18",
doi = "10.18653/v1/2022.nlp4pi-1.18",
pages = "149--156",
abstract = "The various potential of children can be limited by language delay or language impairments. However, there are many instances where parents are unaware of the child{'}s condition and do not obtain appropriate treatment as a result. Additionally, experts collecting children{'}s utterance to establish norms of language tests and evaluating children{'}s language development level takes a significant amount of time and work. To address these issues, dependable automated screening tools are required. In this paper, we used pretrained LM to assist experts in quickly and objectively screening the language development level of children. Here, evaluating the language development level is to ensure that the child has the appropriate language abilities for his or her age, which is the same as the child{'}s age. To do this, we analyzed the utterances of children according to age. Based on these findings, we use the standard deviations of the pretrained LM{'}s probability as a score for children to screen their language development level. The experiment results showed very strong correlations between our proposed method and the Korean language test REVT (REVT-R, REVT-E), with Pearson correlation coefficient of 0.9888 and 0.9892, respectively.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="oh-etal-2022-applicability">
<titleInfo>
<title>Applicability of Pretrained Language Models: Automatic Screening for Children’s Language Development Level</title>
</titleInfo>
<name type="personal">
<namePart type="given">Byoung-doo</namePart>
<namePart type="family">Oh</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Yoon-koung</namePart>
<namePart type="family">Lee</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Yu-seop</namePart>
<namePart type="family">Kim</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2022-12</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Second Workshop on NLP for Positive Impact (NLP4PI)</title>
</titleInfo>
<name type="personal">
<namePart type="given">Laura</namePart>
<namePart type="family">Biester</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Dorottya</namePart>
<namePart type="family">Demszky</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Zhijing</namePart>
<namePart type="family">Jin</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Mrinmaya</namePart>
<namePart type="family">Sachan</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Joel</namePart>
<namePart type="family">Tetreault</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Steven</namePart>
<namePart type="family">Wilson</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Lu</namePart>
<namePart type="family">Xiao</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jieyu</namePart>
<namePart type="family">Zhao</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Abu Dhabi, United Arab Emirates (Hybrid)</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>The various potential of children can be limited by language delay or language impairments. However, there are many instances where parents are unaware of the child’s condition and do not obtain appropriate treatment as a result. Additionally, experts collecting children’s utterance to establish norms of language tests and evaluating children’s language development level takes a significant amount of time and work. To address these issues, dependable automated screening tools are required. In this paper, we used pretrained LM to assist experts in quickly and objectively screening the language development level of children. Here, evaluating the language development level is to ensure that the child has the appropriate language abilities for his or her age, which is the same as the child’s age. To do this, we analyzed the utterances of children according to age. Based on these findings, we use the standard deviations of the pretrained LM’s probability as a score for children to screen their language development level. The experiment results showed very strong correlations between our proposed method and the Korean language test REVT (REVT-R, REVT-E), with Pearson correlation coefficient of 0.9888 and 0.9892, respectively.</abstract>
<identifier type="citekey">oh-etal-2022-applicability</identifier>
<identifier type="doi">10.18653/v1/2022.nlp4pi-1.18</identifier>
<location>
<url>https://aclanthology.org/2022.nlp4pi-1.18</url>
</location>
<part>
<date>2022-12</date>
<extent unit="page">
<start>149</start>
<end>156</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Applicability of Pretrained Language Models: Automatic Screening for Children’s Language Development Level
%A Oh, Byoung-doo
%A Lee, Yoon-koung
%A Kim, Yu-seop
%Y Biester, Laura
%Y Demszky, Dorottya
%Y Jin, Zhijing
%Y Sachan, Mrinmaya
%Y Tetreault, Joel
%Y Wilson, Steven
%Y Xiao, Lu
%Y Zhao, Jieyu
%S Proceedings of the Second Workshop on NLP for Positive Impact (NLP4PI)
%D 2022
%8 December
%I Association for Computational Linguistics
%C Abu Dhabi, United Arab Emirates (Hybrid)
%F oh-etal-2022-applicability
%X The various potential of children can be limited by language delay or language impairments. However, there are many instances where parents are unaware of the child’s condition and do not obtain appropriate treatment as a result. Additionally, experts collecting children’s utterance to establish norms of language tests and evaluating children’s language development level takes a significant amount of time and work. To address these issues, dependable automated screening tools are required. In this paper, we used pretrained LM to assist experts in quickly and objectively screening the language development level of children. Here, evaluating the language development level is to ensure that the child has the appropriate language abilities for his or her age, which is the same as the child’s age. To do this, we analyzed the utterances of children according to age. Based on these findings, we use the standard deviations of the pretrained LM’s probability as a score for children to screen their language development level. The experiment results showed very strong correlations between our proposed method and the Korean language test REVT (REVT-R, REVT-E), with Pearson correlation coefficient of 0.9888 and 0.9892, respectively.
%R 10.18653/v1/2022.nlp4pi-1.18
%U https://aclanthology.org/2022.nlp4pi-1.18
%U https://doi.org/10.18653/v1/2022.nlp4pi-1.18
%P 149-156
Markdown (Informal)
[Applicability of Pretrained Language Models: Automatic Screening for Children’s Language Development Level](https://aclanthology.org/2022.nlp4pi-1.18) (Oh et al., NLP4PI 2022)
ACL