Injecting structural hints: Using language models to study inductive biases in language learning

被引:0
|
作者
Papadimitriou, Isabel [1 ]
Jurafsky, Dan [1 ]
机构
[1] Stanford Univ, Comp Sci Dept, Stanford, CA 94305 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Both humans and large language models are able to learn language without explicit structural supervision. What inductive biases make this learning possible? We address this fundamental cognitive question by leveraging transformer language models: we inject inductive bias into language models by pretraining on formally-structured data, and then evaluate the biased learners' ability to learn typologicallydiverse natural languages. Our experimental setup creates a testbed for hypotheses about inductive bias in human language learning. We investigate the effect of injecting models with three types of inductive bias: 1) recursive, hierarchical processing, 2) crossing token-token relationships that can't be modeled by contextfree grammars, and 3) a Zipfian power-law vocabulary distribution. We show that noncontext-free relationships form the best inductive biases. Our study leverages the capabilities of transformer models to run controlled language learning experiments that are not possible to run on humans, and surfaces hypotheses about the structures that facilitate language learning in both humans and machines.
引用
收藏
页码:8402 / 8413
页数:12
相关论文
共 50 条
  • [1] Learning with Enriched Inductive Biases for Vision-Language Models
    Yang, Lingxiao
    Zhang, Ru-Yuan
    Chen, Qi
    Xie, Xiaohua
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025,
  • [2] Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
    Sartran, Laurent
    Barrett, Samuel
    Kuncoro, Adhiguna
    Stanojevic, Milos
    Blunsom, Phil
    Dyer, Chris
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 1423 - 1439
  • [3] Using Natural Sentences for Understanding Biases in Language Models
    Alnegheimish, Sarah
    Guo, Alicia
    Sun, Yi
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 2824 - 2830
  • [4] Inductive inference and language learning
    Zeugmann, Thomas
    THEORY AND APPLICATIONS OF MODELS OF COMPUTATION, PROCEEDINGS, 2006, 3959 : 464 - 473
  • [5] Study Hints for Language Students
    Moulton, William G.
    MODERN LANGUAGE JOURNAL, 1952, 36 (06): : 259 - 264
  • [6] Using natural language and program abstractions to instill human inductive biases in machines
    Kumar, Sreejan
    Correa, Carlos G.
    Dasgupta, Ishita
    Marjieh, Raja
    Hu, Michael Y.
    Hawkins, Robert D.
    Daw, Nathaniel D.
    Cohen, Jonathan D.
    Narasimhan, Karthik
    Griffiths, Thomas L.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [7] Confirmation and Specificity Biases in Large Language Models: An Explorative Study
    O'Leary, Daniel E.
    IEEE INTELLIGENT SYSTEMS, 2025, 40 (01) : 63 - 68
  • [8] Injecting Numerical Reasoning Skills into Language Models
    Geva, Mor
    Gupta, Ankit
    Beran, Jonathan
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 946 - 958
  • [9] A cognitive modeling approach to learning and using reference biases in language
    Toth, Abigail G. G.
    Hendriks, Petra
    Taatgen, Niels A. A.
    van Rij, Jacolien
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2022, 5
  • [10] Structural Biases in Phonology: Infant and Adult Evidence from Artificial Language Learning
    Bergelson, Elika
    Idsardi, William J.
    PROCEEDINGS OF THE 33RD ANNUAL BOSTON UNIVERSITY CONFERENCE ON LANGUAGE DEVELOPMENT, VOLS 1 AND 2, 2009, : 85 - +