skip to content

Department of Computer Science and Technology

Date: 
Tuesday, 27 January, 2026 - 16:00 to 17:00
Speaker: 
Xiyuan Zhang, Amazon Web Services
Venue: 
Online

Foundation models are transforming structured data learning much like large language models did for text. In this talk, I will present our new foundation models, Mitra and Chronos-2, which demonstrate how synthetic pretraining and in-context learning (ICL) enable models to generalize across diverse tabular and time-series tasks without task-specific training. Mitra curates a principled mixture of synthetic priors to achieve state-of-the-art performance on tabular classification and regression, while Chronos-2 introduces group attention to unify univariate, multivariate, and covariate-informed forecasting. Together, they illustrate a new paradigm where the design of synthetic data priors and ICL mechanisms, rather than per-task fine-tuning, drives generalization and scalability across structured domains.

Bio: Xiyuan Zhang is an Applied Scientist at Amazon Web Services working on machine learning for structured data (time series, tabular), especially on pre-training and multimodal analysis. She is the lead author of Mitra, the most downloaded tabular foundation model on HuggingFace, and co-author of Chronos, the most downloaded time series foundation model on HuggingFace. Xiyuan earned her PhD in Computer Science from the University of California, San Diego. She is a recipient of the Qualcomm Innovation Fellowship and has been recognized as a Cyber-Physical-System (CPS) Rising Star.

https://cam-ac-uk.zoom.us/j/88514183232?pwd=paFmBh7F64OIKIOFoBjQu7Fut65qB6.1
Seminar series: 
Mobile and Wearable Health Seminar Series