Karen Ullrich (s/h) ✈️ COLM
@karen-ullrich.bsky.social
4.2K followers 120 following 26 posts
Research scientist at FAIR NY ❤️ LLMs + Information Theory. Previously, PhD at UoAmsterdam, intern at DeepMind + MSRC.
Posts Media Videos Starter Packs
Pinned
karen-ullrich.bsky.social
#Tokenization is undeniably a key player in the success story of #LLMs but we poorly understand why.
I want to highlight progress we made in understanding the role of tokenization, developing the core incidents and mitigating its problems. 🧵👇
karen-ullrich.bsky.social
Y’all, I am at #COLM this week, very excited to learn, and meet old and new friends. Please reach out on Whova!
karen-ullrich.bsky.social
Plus, we generate importance maps showing where in the transformer the concept is encoded — providing interpretable insights into model internals.
karen-ullrich.bsky.social
SAMI: Diminishes or amplifies these modules to control the concept's influence

With SAMI, we can scale the importance of these modules — either amplifying or suppressing specific concepts.
karen-ullrich.bsky.social
SAMD: Finds the attention heads most correlated with a concept

Using SAMD, we find that only a few attention heads are crucial for a wide range of concepts—confirming the sparse, modular nature of knowledge in transformers.
karen-ullrich.bsky.social
How would you make an LLM "forget" the concept of dog — or any other arbitrary concept? 🐶❓

We introduce SAMD & SAMI — a novel, concept-agnostic approach to identify and manipulate attention modules in transformers.
karen-ullrich.bsky.social
Aligned Multi-Objective Optimization (A-🐮) has been accepted at #ICML2025! 🎉
We explore optimization scenarios where objectives align rather than conflict, introducing new scalable algorithms with theoretical guarantees. #MachineLearning #AI #Optimization
karen-ullrich.bsky.social
🎉🎉 Our paper just got accepted to #ICLR2025! 🎉🎉

Byte-level LLMs without training and guaranteed performance? Curious how? Dive into our work! 📚✨

Paper: arxiv.org/abs/2410.09303
Github: github.com/facebookrese...
Screenshot of arxiv paper "EXACT BYTE-LEVEL PROBABILITIES FROM TOKENIZED LANGUAGE MODELS FOR FIM-TASKS AND MODEL ENSEMBLES."
karen-ullrich.bsky.social
Thursday is busy:
9-11am I will be at the Meta AI Booth
12.30-2pm
Mission Impossible: A Statistical Perspective on Jailbreaking LLMs (neurips.cc/virtual/2024...)
OR
End-To-End Causal Effect Estimation from Unstructured Natural Language Data (neurips.cc/virtual/2024...)
NeurIPS Poster Mission Impossible: A Statistical Perspective on Jailbreaking LLMsNeurIPS 2024
neurips.cc
karen-ullrich.bsky.social
Starting with Fei-Fei Li’s talk 2.30, after that I will mostly be meeting people and wonder the poster sessions.
karen-ullrich.bsky.social
Folks, I am posting my NeurIPS schedule daily in hopes to see folks, thanks @tkipf.bsky.social for the idea ;)

11-12.30 WiML round tables
1.30-4 Beyond Decoding, Tutorial
karen-ullrich.bsky.social
I will be at #Neurips2024 next week to talk about these two papers and host a workshop on #NeuralCompression.
karen-ullrich.bsky.social
🎉 Exciting News! 🎉
Two papers have been accepted at #NeurIPS2024 ! 🙌🏼 These papers are the first outcomes of my growing focus on LLMs. 🍾 Cheers to Nikita Dhawan and Jingtong Su + all involved collaborators: @cmaddis.bsky.social Leo Cotta, Rahul Krishnan, Julia Kempe
karen-ullrich.bsky.social
next one on the list is Yury Polyanskiy's "Information Theory: From Coding to Learning" which will hopefully hit the shelfs in February... can not wait
karen-ullrich.bsky.social
Pro-tip: Use massive black Friday deals at scientific publishing houses to for example buy a copy of @jmtomczak.bsky.social
book on generative modeling (long overdue)
karen-ullrich.bsky.social
What do you think do we need to sharpen our understanding of tokenization? Or will we soon be rid of it by developing models such as "MegaByte" by
Yu et al?
And add more paper to the threat!
karen-ullrich.bsky.social
Phan et al, found a method to mitigate some of the tokenization problems Karpathy mentioned by projecting tokens into byte space. The key to their method is to develop a map between statistically equivalent token and byte-level models.
karen-ullrich.bsky.social
In "The Foundations of Tokenization:
Statistical and Computational Concerns", Gastaldi et al. try to make first steps towards defining what a tokenizer should be and define properties it ought to have.
karen-ullrich.bsky.social
In "Toward a Theory of Tokenization in LLMs" Rajaraman et al., the authors discuss why we can think of tokenization to cause lower perplexity/ a better entropy bound.
karen-ullrich.bsky.social
A must watch entry point is @karpathy.bsky.social hy's "Let's build the GPT Tokenizer" video, where he discusses some tokenization problems.
karen-ullrich.bsky.social
#Tokenization is undeniably a key player in the success story of #LLMs but we poorly understand why.
I want to highlight progress we made in understanding the role of tokenization, developing the core incidents and mitigating its problems. 🧵👇
karen-ullrich.bsky.social
🚨 Internship Opportunity at FAIR NY 🚨

I got one PhD internship position available for 2025!

Interested in exploring the intersection of information theory, probabilistic reasoning, and LLMs?

📩 Send me a DM with your CV, website, and GScholar profile by October 14th.
karen-ullrich.bsky.social
🎉 Exciting News! 🎉
Two papers have been accepted at #NeurIPS2024 ! 🙌🏼 These papers are the first outcomes of my growing focus on LLMs. 🍾 Cheers to Nikita Dhawan and Jingtong Su + all involved collaborators: @cmaddis.bsky.social Leo Cotta, Rahul Krishnan, Julia Kempe