prof_pic.jpg

Benjamin Minixhofer

b{lastname}@gmail.com

I am a third-year PhD student at the Language Technology Lab of the University of Cambridge. My research focuses on:

  • Maximally resource-efficient synthesis of insights and models: I believe retrofitting is a highly promising, underrated area of research that may allow gaining architectural insights and creating models at a small fraction of the cost of current common practices. I also believe in the utility of creating effective ultra-efficient task-specialized models.
  • Improving modularity and malleability of LLMs: As ecosystems emerge around LLMs, there is a risk of these ecosystems remaining mutually incompatible. I aim to improve the modularity (degree of compartmentalization) and malleability (ease of adaptation) of LLMs to remove barriers across LLM ecosystems, aiming to enable new ways of information interchange and collaboration across LLMs.
  • Making units of computation (a.k.a. tokens) adaptive: A fundamental restriction of current LLMs is that the tokenization remains largely fixed after the one-time decision how to tokenize. I work on ways to make tokenization adaptive, such as Zero-Shot Tokenizer Transfer and Cross-Tokenizer Distillation, with the primary goal of improving inference efficiency and equity across languages.

I’ve been fortunate to gain research experience at Ai2, Google DeepMind, Cohere, H2O.ai, Huawei Noah’s Ark Lab (2x), Johannes Kepler University Linz (where I also did my undergrad) and through competing on Kaggle.

If you’d like to chat about any of these topics (or about 🐱⛵🥾), drop me an email!

News

Jan 21, 2026 I wrote a (my first!) blog post on Four Ingredients for Successful Language Model Retrofitting.
Dec 19, 2025 Our new preprint Bolmo: Byteifying the Next Generation of Language Models is up on arXiv!
Sep 18, 2025 Universal Cross-Tokenizer Distillation via Approximate Likelihood Matching is accepted at NeurIPS! See you (probably) at EurIPS!
Jul 7, 2025 Started an internship at Ai2 in Seattle over the Summer to work with Luca Soldaini and Valentin Hofmann!
Nov 29, 2024 Talk about «The Past, Present and Future of Tokenization» at the NLIP Seminar in Cambridge. This talk was based on an Invited Lecture at the University of Göttingen in early November. Slides.
Nov 27, 2024 Attended the ELLIS NLP Workshop at Dagstuhl. Some nice photos.
Sep 25, 2024 Zero-Shot Tokenizer Transfer is accepted at NeurIPS 2024. See you in Vancouver! :canada:
Jul 24, 2024 I presented Zero-Shot Tokenizer Transfer at Google DeepMind and Mozilla. Slides.

Selected Publications

  1. Bolmo: Byteifying the Next Generation of Language Models
    Benjamin Minixhofer, Tyler Murray, Tomasz Limisiewicz, Anna Korhonen, Luke Zettlemoyer, Noah A. Smith, Edoardo M. Ponti, Luca Soldaini, and Valentin Hofmann
    In arXiv preprint, Dec 2025
  2. Cross-Tokenizer Distillation via Approximate Likelihood Matching
    Benjamin Minixhofer, Ivan Vulić, and Edoardo Maria Ponti
    In The Thirty-ninth Annual Conference on Neural Information Processing Systems, Dec 2025
  3. Zero-Shot Tokenizer Transfer
    Benjamin Minixhofer, Edoardo Ponti, and Ivan Vulić
    In The Thirty-eighth Annual Conference on Neural Information Processing Systems, Dec 2024
  4. Segment Any Text: A Universal Approach for Robust, Efficient and Adaptable Sentence Segmentation
    Markus Frohmann, Igor Sterner, Ivan Vulić*, Benjamin Minixhofer*, and Markus Schedl*
    In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Nov 2024

Invited Talks

  • University of Toronto (2026) — Tokenizer Transfer
  • University College London (2026) — Tokenization in Modern LLMs (Guest Lecture)
  • University of Edinburgh (2025) — Tokenization in Modern LLMs (Guest Lecture)
  • University of Cambridge (2024) — The Past, Present, and Future of Tokenization
  • University of Göttingen (2024) — The Past, Present, and Future of Tokenization (Guest Lecture)
  • Google DeepMind (2024) — Zero-Shot Tokenizer Transfer
  • Mozilla (2024) — Zero-Shot Tokenizer Transfer