Séminaire de Recherche en Linguistique

Ce séminaire reçoit des conférenciers invités spécialisés dans différents domaines de la linguistique. Les membres du Département, les étudiants et les personnes externes intéressées sont tous cordialement invités.

Description du séminaire Print

Titre Language Models and Human Language Acquisition
Conférencier Alex Warstadt ( ETH Zürich)
Date mardi 20 février 2024
Heure 12h15
Salle L208 (Bâtiment Candolle)
Description

Children’s remarkable ability to learn language has been an object of fascination in science for millennia. But in just the last few years, neural language models (LMs) have also proven to be incredibly adept at learning human language. I argue that—under the right circumstances—we can use LMs to test key hypotheses about language acquisition that have been difficult or impossible to evaluate with human subjects. As a proof-of-concept, I obtain experimental evidence for the controversial claim that language models can acquire structural rules such as subject-auxiliary inversion in English without the aid of direct evidence in the input. Nonetheless, obstacles remain in the way of using LMs as models of human language acquisition. LMs are extremely inefficient learners compared to humans, requiring hundreds or thousands of times more data to reach human-like grammatical generalizations. I present results from the first BabyLM Challenge, a shared task challenging researchers to train LMs with constrained data inspired by the input to children. From over 30 submissions, including my own work, we uncover promising architectural and data preprocessing advances which improve data efficiency, and we explore multimodality and interaction as methods to increase developmental plausibility of language models.


 

 

   
Document(s) joint(s) -