Tracking Child Language Development With Neural Network Language Models

被引:1
|
作者
Sagae, Kenji [1 ]
机构
[1] Univ Calif Davis, Dept Linguist, Davis, CA 95616 USA
来源
FRONTIERS IN PSYCHOLOGY | 2021年 / 12卷
关键词
child language assessment; natural language processing; computational linguistics; language model; IPSyn; neural network; PRODUCTIVE SYNTAX; INDEX;
D O I
10.3389/fpsyg.2021.674402
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Recent work on the application of neural networks to language modeling has shown that models based on certain neural architectures can capture syntactic information from utterances and sentences even when not given an explicitly syntactic objective. We examine whether a fully data-driven model of language development that uses a recurrent neural network encoder for utterances can track how child language utterances change over the course of language development in a way that is comparable to what is achieved using established language assessment metrics that use language-specific information carefully designed by experts. Given only transcripts of child language utterances from the CHILDES Database and no pre-specified information about language, our model captures not just the structural characteristics of child language utterances, but how these structures reflect language development over time. We establish an evaluation methodology with which we can examine how well our model tracks language development compared to three known approaches: Mean Length of Utterance, the Developmental Sentence Score, and the Index of Productive Syntax. We discuss the applicability of our model to data-driven assessment of child language development, including how a fully data-driven approach supports the possibility of increased research in multilingual and cross-lingual issues.
引用
收藏
页数:14
相关论文
共 50 条