Publications

2025

  1. Manuscript
    A Distributional Perspective on Word Learning in Neural Language Models
    Filippo Ficarra, Ryan Cotterell, and Alex Warstadt
    2025
  2. Manuscript
    Can Language Models Learn Typologically Implausible Languages?
    Tianyang Xu, Tatsuki Kuribayashi, Yohei Oseki, and 2 more authors
    2025

2024

  1. TACL (to appear)
    Investigating Critical Period Effects in Language Acquisition through Neural Language Models
    Ionut Constantinescu, Tiago Pimentel, Ryan Cotterell, and 1 more author
    Transactions of the Association for Computational Linguistics, 2024
  2. COLING
    Automatic annotation of grammaticality in child-caregiver conversations
    Mitja Nikolaus, Abhishek Agrawal, Petros Kaklamanis, and 2 more authors
    In Proceedings of the 2024 joint international conference on computational linguistics, language resources and evaluation (LREC-COLING 2024), May 2024
  3. Under Review
    Bigger is not always better: The importance of human-scale language modeling for psycholinguistics
    Ethan Gotlieb Wilcox, Michael Hu, Aaron Mueller, and 6 more authors
    May 2024
  4. EMNLP
    Surprise! Uniform Information Density Isn’t the Whole Story: Predicting Surprisal Contours in Long-form Discourse
    Eleftheria Tsipidi, Franz Nowak, Ryan Cotterell, and 3 more authors
    In Proceedings of the 2023 conference on empirical methods in natural language processing (EMNLP), Nov 2024

2023

  1. Under Review
    A geometric notion of causal probing
    Clément Guerner, Anej Svete, Tianyu Liu, and 2 more authors
    arXiv preprint arXiv:2307.15054, Nov 2023
  2. ACL
    Generalizing backpropagation for gradient-based interpretability
    Kevin Du, Lucas Torroba Hennigen, Niklas Stoehr, and 2 more authors
    In Proceedings of the 61st annual meeting of the association for computational linguistics (volume 1: Long papers), Jul 2023
  3. BabyLM
    WhisBERT: Multimodal text-audio language modeling on 100M words
    Lukas Wolf, Klemen Kotar, Greta Tuckute, and 4 more authors
    In Proceedings of the BabyLM challenge at the 27th conference on computational natural language learning, Dec 2023
  4. BabyLM
    Findings of the BabyLM challenge: Sample-efficient pretraining on developmentally plausible corpora
    Alex Warstadt, Aaron Mueller, Leshem Choshen, and 8 more authors
    In Proceedings of the BabyLM challenge at the 27th conference on computational natural language learning, Dec 2023
  5. BabyLM
    Acquiring linguistic knowledge from multimodal input
    Theodor Amariucai, and Alexander Scott Warstadt
    In Proceedings of the BabyLM challenge at the 27th conference on computational natural language learning, Dec 2023
  6. BabyLM
    Proceedings of the BabyLM Challenge at the 27th Conference on Computational Natural Language Learning
    Dec 2023

2022

  1. Book Chapter
    What artificial neural networks can tell us about human language acquisition
    Alex Warstadt, and Samuel R Bowman
    In Algebraic Structures in Natural Language, Dec 2022
  2. Diss. Chapter
    The Role of Indirect Evidence in Grammar Learning: Investigations with Causal Manipulations of the Learning Environment
    Alex Warstadt
    In Artificial neural networks as models of human language acquisition, Dec 2022

2021

  1. ACL
    When Do You Need Billions of Words of Pretraining Data?
    Yian Zhang, Alex Warstadt, Xiaocheng Li, and 1 more author
    In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Aug 2021