Ai2
banner
ai2.bsky.social
Ai2
@ai2.bsky.social
Breakthrough AI to solve the world's biggest problems.

› Join us: http://allenai.org/careers
› Get our newsletter: https://share.hsforms.com/1uJkWs5aDRHWhiky3aHooIg3ioxm
Pinned
Introducing Ai2 Open Coding Agents—starting with SERA, our first-ever coding models. Fast, accessible agents (8B–32B) that adapt to any repo, including private codebases. Train a powerful specialized agent for as little as ~$400, & it works with Claude Code out of the box. 🧵
Reposted by Ai2
Many want to use AI to accelerate science, and utilizing it to explore the growing tsunami of research articles is getting lots of attention. Measuring the quality of AI answers to questions about science is a challenge. @science.org www.science.org/content/arti...
Open-source AI program can answer science questions better than humans
Developed by and for academics, OpenScholar aims to improve searches of the ballooning scientific literature
www.science.org
February 4, 2026 at 6:52 PM
Our OpenScholar paper is now in @nature.com 🎉

OpenScholar is an open-source model for synthesizing scientific research—with citations as accurate as human experts. 🧵
February 4, 2026 at 4:21 PM
Since launching Open Coding Agents, it's been exciting to see how quickly the community has adopted them. Today we're releasing SERA-14B – a new 14B-parameter coding model – plus a major refresh of our open training datasets. 🧵
February 3, 2026 at 5:39 PM
Introducing Theorizer: Turning thousands of papers into scientific laws 📚➡️📜

Most automated discovery systems focus on experimentation. Theorizer tackles the other half of science: theory building—compressing scattered findings into structured, testable claims. 🧵
January 28, 2026 at 6:37 PM
Reposted by Ai2
Here's just one of the cool apps you can vibe-code with SERA, our new agentic coding model! I was lucky enough to get my hands on it early and it's quite capable via Claude Code. Give it a go today!
January 27, 2026 at 8:29 PM
Introducing Ai2 Open Coding Agents—starting with SERA, our first-ever coding models. Fast, accessible agents (8B–32B) that adapt to any repo, including private codebases. Train a powerful specialized agent for as little as ~$400, & it works with Claude Code out of the box. 🧵
January 27, 2026 at 4:13 PM
Molmo 2 (8B) is now available via @hf.co Inference Providers, courtesy of Public AI.

State-of-the-art video understanding with pointing, counting, & multi-frame reasoning. Track objects through scenes and identify where + when events occur. 🧵
January 26, 2026 at 5:16 PM
Introducing HiRO-ACE: an AI framework that makes highly detailed climate simulations dramatically more accessible. It generates decades of high-resolution precipitation data for any region in a day on a single GPU—no supercomputing cluster required. 🧵
January 21, 2026 at 7:34 PM
"We wanted to provide subject matter experts and communities that have the expertise on the ground with the tools to engage with AI without having to learn AI deeply.” - Ted Schmitt. Thanks @mongabay.com for diving into our new OlmoEarth platform. 📷
January 16, 2026 at 8:55 PM
SciArena update: our Olmo 3.1 32B Instruct scores 963.6 Elo overall at just $0.17/100 calls—ahead of OpenAI’s GPT-OSS-20B. In Engineering, it hits 1039.2 Elo, only 2.5 behind GPT-OSS-120B—a model ~4× its size. 🧵
January 16, 2026 at 5:57 PM
Molmo 2 is now available via API on @openrouter.bsky.social, courtesy of Parasail—free until 1/29.
State-of-the-art video understanding with pointing, counting, and multi-frame reasoning—track objects through scenes & identify where + when events occur.
Open. Apache 2.0. 👇
January 13, 2026 at 5:59 PM
Olmo 3.1 32B Instruct is now on @openrouter.bsky.social, hosted by DeepInfra. Built for real-world use: reliable instruction following & function calling for agentic workflows + research. Fully open & leading benchmark performance, ready to plug into your stack. 👇
January 8, 2026 at 8:00 PM
🆕 New in Asta: multi-turn report generation.
You can now have back-and-forth conversations with Asta, our agentic platform for scientific research, to refine long-form, fully cited reports instead of relying on single-shot prompts.
December 18, 2025 at 4:09 PM
Now you can use our most powerful models via API.
Olmo 3.1 32B Think, our reasoning model for complex problems, is on @openrouter.bsky.social—free through 12/22. And Olmo 3.1 32B Instruct, our flagship chat model with tool use, is available through @hf.co Inference Providers. 👇
December 17, 2025 at 9:02 PM
🎥 Introducing SAGE, an agentic system for long video reasoning on entertainment videos—sports, vlogs, & more. It learns when to skim, zoom in, & answer questions directly. On our SAGE-Bench eval, SAGE with a Molmo 2 (8B)-based orchestrator lifts accuracy from 61.8% → 66.1%. 🧵
December 17, 2025 at 5:57 PM
🎗️ Reminder, our Molmo 2 and Olmo 3 Reddit AMA begins soon at 1pm PST / 4pm EST. www.reddit.com/r/LocalLLaMA...
From the LocalLLaMA community on Reddit
Explore this post and more from the LocalLLaMA community
www.reddit.com
December 16, 2025 at 8:41 PM
Last year Molmo set SOTA on image benchmarks + pioneered image pointing. Millions of downloads later, Molmo 2 brings Molmo’s grounded multimodal capabilities to video 🎥—and leads many open models on challenging industry video benchmarks. 🧵
December 16, 2025 at 4:52 PM
🗓️ Tue Dec 16, 1–2pm PT: AMA with researchers + engineers from our Olmo & Molmo teams, hosted by r/LocalLLaMA.
💬 Ask your questions now—we’ll start answering when the AMA begins!
December 15, 2025 at 10:25 PM
Introducing Bolmo, a new family of byte-level language models built by "byteifying" our open Olmo 3—and to our knowledge, the first fully open byte-level LM to match or surpass SOTA subword models across a wide range of tasks. 🧵
December 15, 2025 at 5:19 PM
🧠 Introducing NeuroDiscoveryBench. Built with @alleninstitute.org, it’s the first benchmark for evaluating AI systems like our Asta DataVoyager agent on neuroscience data. The benchmark tests whether AI can truly extract insights from complex brain datasets.
December 12, 2025 at 8:41 PM
Olmo 3.1 is here. We extended our strongest RL run and scaled our instruct recipe to 32B—releasing Olmo 3.1 Think 32B & Olmo 3.1 Instruct 32B, our most capable models yet. 🧵
December 12, 2025 at 5:14 PM
Update: DataVoyager, which we launched in Preview early this fall, is now available in Asta. 🎉
You can upload real datasets, ask complex research questions in natural language, & get back reproducible answers + visualizations. 🔍📊
December 8, 2025 at 8:47 PM
We're at #NeurIPS2025 with papers, posters, workshops, fireside chats, & talks across the conference. Come learn about our latest research + see live demos!
December 2, 2025 at 6:05 PM
SciArena leaderboard update! 🔬
We've added new frontier models – including GPT-5.1 and Gemini 3 Pro Preview – to our arena for scientific literature tasks. The new rankings: o3 holds #1, Gemini 3 Pro Preview lands at #2, Claude Opus 4.1 sits at #3, GPT-5 at #4, & GPT-5.1 debuts at #5. 🧵
December 1, 2025 at 8:24 PM
Olmo 3 is now available through @hf.co Inference Providers, thanks to Public AI! 🎉
This means you can run our fully open 7B and 32B models — including Think and Instruct variants — via serverless API with no infrastructure to manage.
November 28, 2025 at 4:50 PM