skip to content

Department of Computer Science and Technology

Date: 
Friday, 6 June, 2025 - 15:00 to 16:00
Speaker: 
Catherine Arnett and Tyler Chang (EleutherAI and UC San Diego)
Venue: 
ONLINE ONLY. Here is the Zoom link: https://cam-ac-uk.zoom.us/j/4751389294?pwd=Z2ZOSDk0eG1wZldVWG1GVVhrTzFIZz09

NOTE THE UNUSUAL TIME FOR THIS SEMINAR

Language models work well for a small number of languages. For the other languages, the best existing language model is likely multilingual, still with the vast majority of the training data coming from English and a few "priority" languages. We show that in many cases, multilinguality leads to worse performance across many languages due to limited model capacity. We then train a suite of over 1,000 monolingual models for 350 languages, finding that these models can outperform multilingual models over ten times their size. However, multilinguality can also be a blessing: we train a small number of controlled bilingual models in order to study how crosslingual transfer happens. We aim to better understand transfer learning in order to better leverage multilinguality to improve language model performance for all languages.

Seminar series: 
NLIP Seminar Series

Upcoming seminars