Benjamin Minixhofer
b{lastname}@gmail.comI am a third-year PhD student at the Language Technology Lab of the University of Cambridge. My research focuses on:
- Maximally resource-efficient synthesis of insights and models: I believe retrofitting is a highly promising, underrated area of research that may allow gaining architectural insights and creating models at a small fraction of the cost of current common practices. I also believe in the utility of creating effective ultra-efficient task-specialized models.
- Improving modularity and malleability of LLMs: As ecosystems emerge around LLMs, there is a risk of these ecosystems remaining mutually incompatible. I aim to improve the modularity (degree of compartmentalization) and malleability (ease of adaptation) of LLMs to remove barriers across LLM ecosystems, aiming to enable new ways of information interchange and collaboration across LLMs.
- Making units of computation (a.k.a. tokens) adaptive: A fundamental restriction of current LLMs is that the tokenization remains largely fixed after the one-time decision how to tokenize. I work on ways to make tokenization adaptive, such as Zero-Shot Tokenizer Transfer and Cross-Tokenizer Distillation, with the primary goal of improving inference efficiency and equity across languages.
I’ve been fortunate to gain research experience at Ai2, Google DeepMind, Cohere, H2O.ai, Huawei Noah’s Ark Lab (2x), Johannes Kepler University Linz (where I also did my undergrad) and through competing on Kaggle.
If you’d like to chat about any of these topics (or about 🐱⛵🥾), drop me an email!
News
| Jan 21, 2026 | I wrote a (my first!) blog post on Four Ingredients for Successful Language Model Retrofitting. |
|---|---|
| Dec 19, 2025 | Our new preprint Bolmo: Byteifying the Next Generation of Language Models is up on arXiv! |
| Sep 18, 2025 | Universal Cross-Tokenizer Distillation via Approximate Likelihood Matching is accepted at NeurIPS! See you (probably) at EurIPS! |
| Jul 7, 2025 | Started an internship at Ai2 in Seattle over the Summer to work with Luca Soldaini and Valentin Hofmann! |
| Nov 29, 2024 | Talk about «The Past, Present and Future of Tokenization» at the NLIP Seminar in Cambridge. This talk was based on an Invited Lecture at the University of Göttingen in early November. Slides. |
| Nov 27, 2024 | Attended the ELLIS NLP Workshop at Dagstuhl. Some nice photos. |
| Sep 25, 2024 |
Zero-Shot Tokenizer Transfer is accepted at NeurIPS 2024. See you in Vancouver! |
| Jul 24, 2024 | I presented Zero-Shot Tokenizer Transfer at Google DeepMind and Mozilla. Slides. |
Selected Publications
-
Segment Any Text: A Universal Approach for Robust, Efficient and Adaptable Sentence SegmentationIn Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Nov 2024
Invited Talks
- University of Toronto (2026) — Tokenizer Transfer
- University College London (2026) — Tokenization in Modern LLMs (Guest Lecture)
- University of Edinburgh (2025) — Tokenization in Modern LLMs (Guest Lecture)
- University of Cambridge (2024) — The Past, Present, and Future of Tokenization
- University of Göttingen (2024) — The Past, Present, and Future of Tokenization (Guest Lecture)
- Google DeepMind (2024) — Zero-Shot Tokenizer Transfer
- Mozilla (2024) — Zero-Shot Tokenizer Transfer