Do BabyLMs Wanna Learn Wanna Contraction? On the Learnability without Language-Specific Bias

ACL ARR 2026 January Submission8262 Authors

06 Jan 2026 (modified: 20 Mar 2026)ACL ARR 2026 January SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: BabyLM, wanna contraction, learnability, CHILDES, surprisal
Abstract: This study investigates whether the grammatical constraints on wanna contraction—a phenomenon traditionally cited as evidence for innate linguistic knowledge—can be learned via BabyLMs, which are designed to reflect cognitively plausible learning conditions. Two datasets were constructed from the CHILDES corpus, varying in embedded verb frequency (high vs. low) and grammaticality, and contrasting grammatical instances (object extraction contexts) with ungrammatical ones (subject extraction contexts) of wanna contractions. Using surprisal as a metric, we evaluated 24 BabyLMs from the 2024 BabyLM Challenge alongside four standard pretrained models, including BERT and GPT-2. While the pretrained models performed with near-perfect consistency, the BabyLMs showed limited but negligible sensitivity, particularly those trained on larger datasets and tested on high-frequency wanna instances. In particular, only encoder-based BabyLMs captured the grammatical constraint, with babylm24_MLSM exhibiting consistent performance. Nonetheless, our findings provide evidence for limited and conditional learnability of wanna contraction by artificial learners under cognitively realistic input conditions.
Paper Type: Long
Research Area: Linguistic theories, Cognitive Modeling and Psycholinguistics
Research Area Keywords: linguistic theories, cognitive modeling, computational psycholinguistics
Contribution Types: Theory
Languages Studied: English
Submission Number: 8262
Loading