skip to content

Department of Computer Science and Technology

Wednesday, 22 November, 2023 - 15:05 to 15:55
Professor Mamoru Komachi - Research Professor, Tokyo Metropolitan University. Visiting academic at the Department of Computer Science and Technology, University of Cambridge
Lecture Theatre 1, Computer Laboratory, William Gates Building

This talk presents an exploration into Multilingual Masked Language Models (MMLMs) as an emerging asset for cross-lingual transfer learning. The focus will be on introducing the mechanisms and applications that position MMLMs at the forefront of advancing multilingual capabilities in NLP.

We'll dissect the transformer architecture that underpins MMLMs, delve into the masking mechanism, and discuss the transfer learning training that enables these models to understand and generate multilingual text. The synergy between these components is critical for the model's linguistic versatility.

Further, the discussion will pivot to optimizing few-shot learning within the MMLM framework. By strategically annotating challenging instances, we can amplify model performance. I'll present findings on employing zero-shot learning techniques to identify such instances for cross-lingual transfer, which could inform annotation strategies.

Attendees will gain a clear understanding of MMLMs, informed by practical applications such as grammatical error correction and sentiment analysis, potentially stimulating further research in the domain.

Link to join virtually:

This talk is being recorded.

Seminar series: 
Wednesday Seminars

Upcoming seminars