Despite massive investments in training large language models, tokenizers remain a critical but often neglected component with weaknesses that can cause wild hallucinations, bypass safety guardrails, and break downstream applications. This talk will cover:
Our recent research in automatically detecting problematic 'glitch' tokens in any model
Fundamental issues with pretokenizers and their design
Novel approaches to encodings and pretokenization that address some of these problems.
**Speaker Bio**
Sander Land is a researcher at Writer, previously working at Cohere. He completed his PhD at the Department of Computer Science, University of Oxford, before undertaking a postdoc at Biomedical Engineering, King's College London, University of London.