Jiacheng Liu
@liujch1998.bsky.social
87 followers 52 following 16 posts
🎓 PhD student @uwcse @uwnlp. 🛩 Private pilot. Previously: 🧑‍💻 @oculus, 🎓 @IllinoisCS. 📖 🥾 🚴‍♂️ 🎵 ♠️
Posts Media Videos Starter Packs
Pinned
liujch1998.bsky.social
Today we're unveiling OLMoTrace, a tool that enables everyone to understand the outputs of LLMs by connecting to their training data.

We do this on unprecedented scale and in real time: finding matching text between model outputs and 4 trillion training tokens within seconds. ✨
ai2.bsky.social
For years it’s been an open question — how much is a language model learning and synthesizing information, and how much is it just memorizing and reciting?

Introducing OLMoTrace, a new feature in the Ai2 Playground that begins to shed some light. 🔦
Reposted by Jiacheng Liu
ai2.bsky.social
Ai2 @ai2.bsky.social · Jun 30
How can we better understand how models make predictions and which components of a training dataset are shaping their behaviors? In April we introduced OLMoTrace, a feature that lets you trace the outputs of language models back to their full training data in real time. 🧵
Reposted by Jiacheng Liu
ai2.bsky.social
As we’ve been working towards training a new version of OLMo, we wanted to improve our methods for measuring the Critical Batch Size (CBS) of a training run, to unlock greater efficiency. but we found gaps between the methods in the literature and our practical needs for training OLMo. 🧵
Reposted by Jiacheng Liu
saumyamalik.bsky.social
I’m thrilled to share RewardBench 2 📊— We created a new multi-domain reward model evaluation that is substantially harder than RewardBench, we trained and released 70 reward models, and we gained insights about reward modeling benchmarks and downstream performance!
Reposted by Jiacheng Liu
ai2.bsky.social
📢We’re taking your questions now on Reddit for tomorrow’s AMA!

Ask us anything about OLMo, our family of fully-open language models. Our researchers will be on hand to answer them Thursday, May 8 at 8am PST.
Reposted by Jiacheng Liu
ai2.bsky.social
The story of OLMo, our Open Language Model, goes back to February 2023 when a group of researchers gathered at Ai2 and started planning. What if we made a language model with state-of-the-art performance, but we did it completely in the open? 🧵
Reposted by Jiacheng Liu
ai2.bsky.social
We're excited to round out the OLMo 2 family with its smallest member, OLMo 2 1B, surpassing peer models like Gemma 3 1B or Llama 3.2 1B. The 1B model should enable rapid iteration for researchers, more local development, and a more complete picture of how our recipe scales.
A bar graph comparing average performance (10 Tasks) across OLMo 2 1B, SmolLM2 1.7B, Gemma 3 1B, Llama 3.2 1B, and Qwen 2.5 1.5B. The highest performance is 42.7, achieved by OLMo 2 1B.
Reposted by Jiacheng Liu
ai2.bsky.social
Have questions? We’re an open book!

We’re excited to host an AMA to answer your Qs about OLMo, our family of open language models.

🗓️ When: May 8, 8-10 am PT
🌐 Where: r/huggingface
🧠 Why: Gain insights from our expert researchers

Chat soon!
Ask Us Anything about our Open Language Model, OLMo
Reposted by Jiacheng Liu
ai2.bsky.social
Ai2 @ai2.bsky.social · Apr 14
Last week we released OLMoTrace as part of #GoogleCloudNext
"With OLMoTrace, we’re actually bringing accessibility to openness, enabling everybody to start looking into the inner workings of the relationships between the input and output of these models." - Ali Farhadi, Ai2 CEO
Reposted by Jiacheng Liu
natolambert.bsky.social
Ai2 launched a new tool where your responses from OLMo get mapped back to related training data. We're using this actively to improve our post-training data and hope many others will use it for understanding and transparency around leading language models!
Some musings:
Looking at the training data
On building tools where truly open-source models can shrine (OLMo 2 32B Instruct, for today). OLMoTrace lets you poke around.
buff.ly
Reposted by Jiacheng Liu
ai2.bsky.social
Ai2 @ai2.bsky.social · Apr 10
Coming to you live from #GoogleCloudNext Day 2!

📍 Find us at the Vertex AI Model Garden inside the Google Cloud Showcase - try out OLMoTrace, and take a step inside our fully open AI ecosystem.
Lead OLMoTrace researcher Jiacheng Liu at Ai2's Google Cloud Next booth. The entrance to the Vertex AI Model Garden at Google Cloud Next. A QR code leading to the story of Google Cloud and Ai2's partnership sitting near a faux fire pit. Ai2 COO Sophie Lebrecht talks to visitors at Ai2's booth at Google Cloud Next.
Reposted by Jiacheng Liu
ai2.bsky.social
Ai2 @ai2.bsky.social · Apr 10
"OLMoTrace is a breakthrough in AI development, setting a new standard for transparency and trust. We hope it will empower researchers, developers, and users to build with confidence—on models they can understand and trust." - CEO Ali Farhadi at tonight's chat with Karen Dahut #GoogleCloudNext
Ali Farhadi speaking on stage at a fireside chat
liujch1998.bsky.social
OLMoTrace is powered by my previous work infini-gram, with some innovative algorithmic twists. Really proud to turn an academic research project into a real LLM product, it’s been a truly amazing experience.

Check out infini-gram: infini-gram.io
liujch1998.bsky.social
Try OLMoTrace in Ai2 Playground with our OLMo 2 models: playground.allenai.org

If OLMoTrace gives you new insight into how LLMs behave, we’d love you to share your use case! 💡Take a screenshot, post the thread link if you like, and don’t forget to tag
@allen_ai
playground.allenai.org
liujch1998.bsky.social
Today we're unveiling OLMoTrace, a tool that enables everyone to understand the outputs of LLMs by connecting to their training data.

We do this on unprecedented scale and in real time: finding matching text between model outputs and 4 trillion training tokens within seconds. ✨
ai2.bsky.social
For years it’s been an open question — how much is a language model learning and synthesizing information, and how much is it just memorizing and reciting?

Introducing OLMoTrace, a new feature in the Ai2 Playground that begins to shed some light. 🔦
Reposted by Jiacheng Liu
ai2.bsky.social
For years it’s been an open question — how much is a language model learning and synthesizing information, and how much is it just memorizing and reciting?

Introducing OLMoTrace, a new feature in the Ai2 Playground that begins to shed some light. 🔦
Reposted by Jiacheng Liu
ai2.bsky.social
📰Google Cloud moves deeper into open source AI with Ai2 partnership:

“Many were wary of using AI models unless they had full transparency into models’ training data and could customize the models completely. Ai2’s models allow that.”
liujch1998.bsky.social
(4/4) Searching in OLMo 2's training data is now available in both our web interface and the API endpoint.

Plus, OLMo 2 32B Instruct is a very strong model. Let's do real science with it 🧪
liujch1998.bsky.social
(3/4) We know the pain point in LLM research in academia: We don't know what's in the training data of these LLMs (GPT, Llama, etc) and what's not; we can only speculate.

So we made the full training data of OLMo 2 and OLMoE searchable, including pre-training and post-training.
liujch1998.bsky.social
(2/4) Check out the source code of infini-gram here: github.com/liujch1998/infini-gram

If you are new to infini-gram, you might want to start with exploring our web interface infini-gram.io/demo and API endpoint infini-gram.io/api_doc
GitHub - liujch1998/infini-gram
Contribute to liujch1998/infini-gram development by creating an account on GitHub.
github.com
liujch1998.bsky.social
As infini-gram surpasses 500 million API calls, today we're announcing two exciting updates:

1. Infini-gram is now open-source under Apache 2.0!
2. We indexed the training data of OLMo 2 models. Now you can search in the training data of these strong, fully-open LLMs.

🧵 (1/4)
Reposted by Jiacheng Liu
ai2.bsky.social
Stay tuned... Wednesday, at #GoogleCloudNext and online 👀
Reposted by Jiacheng Liu
ai2.bsky.social
Buckle your seatbelt — we've released the OLMo 2 paper to kick off 2025 🔥. Including 50+ pages on 4 crucial components of the LLM development pipeline.
A list of paper authors for 2 OLMo 2 Furious.
Reposted by Jiacheng Liu
kylelo.bsky.social
kicking off 2025 with our OLMo 2 tech report while payin homage to the sequelest of sequels 🫡

🚗 2 OLMo 2 Furious 🔥 is everythin we learned since OLMo 1, with deep dives into:

🚖 stable pretrain recipe
🚔 lr anneal 🤝 data curricula 🤝 soups
🚘 tulu post-train recipe
🚜 compute infra setup

👇🧵
liujch1998.bsky.social
Yes we’ve read your paper and there’s so many interesting findings! Let’s grab coffee at Neurips