Martin Wattenberg
@wattenberg.bsky.social
7.6K followers 1.2K following 390 posts
Human/AI interaction. ML interpretability. Visualization as design, science, art. Professor at Harvard, and part-time at Google DeepMind.
Posts Media Videos Starter Packs
Reposted by Martin Wattenberg
vrli.bsky.social
Charts and graphs help people analyze data, but can they also help AI?

In a new paper, we provide initial evidence that it does! GPT 4.1 and Claude 3.5 describe three synthetic datasets more precisely and accurately when raw data is accompanied by a scatter plot. Read more in🧵!
Reposted by Martin Wattenberg
zittrain.bsky.social
AI is often thought of as a black box -- no way to know what's going on inside. That's changing in eye-opening ways. Researchers are finding "beliefs" models are forming as they converse, and how those beliefs correlate to what the models say and how they say it.

www.theatlantic.com/technology/a...
What AI Thinks It Knows About You
What happens when people can see what assumptions a large language model is making about them?
www.theatlantic.com
Reposted by Martin Wattenberg
babynames.bsky.social
The interactive NameGrapher is updated with 2024 baby name popularity stats! Come explore--and marvel that Oliver and Olivia have converged namerology.com/baby-name-gr...
Historical popularity chart showing the popularity of Oliver rising to meet the previously much greater popularity of Olivia
wattenberg.bsky.social
A wonderful visualization for those of us obsessed by sunlight and geography!
oliviafvane.bsky.social
This map shows the hour of sunrise globally through the year. It reveals time zones following national and, sometimes, regional boundaries, and slicing through the oceans.
wattenberg.bsky.social
An incredibly rich, detailed view of neural net internals! There are so many insights in these papers. And the visualizations of "addition circuit" features are just plain cool!
colah.bsky.social
Can we understand the mechanisms of a frontier AI model?

📝 Blog post: www.anthropic.com/research/tra...
🧪 "Biology" paper: transformer-circuits.pub/2025/attribu...
⚙️ Methods paper: transformer-circuits.pub/2025/attribu...

Featuring basic multi-step reasoning, planning, introspection and more!
On the Biology of a Large Language Model
transformer-circuits.pub
wattenberg.bsky.social
Great news, congrats! And glad you’ll still be in the neighborhood!
wattenberg.bsky.social
I'd be curious about advice on teaching non-coders how to test programs they've written with AI. I'm not thinking unit tests so much as things like making sure you can drill down for verifiable details in a visualization—basic practices that are good on their own, but also help catch errors.
wattenberg.bsky.social
Now that we have vibe coding, we need vibe testing!
wattenberg.bsky.social
Oh, that looks super relevant and fascinating, reading through it now...
wattenberg.bsky.social
Ha! I think (!) that for me, the word "calculate" connotes narrow precision and correctness, whereas "think" is more expansive but also implies more fuzziness and the possibility of being wrong. That said, your observation does give me pause!
wattenberg.bsky.social
Interesting question! I haven't calculated this, but @yidachen.bsky.social might know
wattenberg.bsky.social
This is a common pattern, but we're also seeing some others! Here are similar views for multiple-choice abstract algebra questions (green is the correct answer; other colors are incorrect answers) You can see many more at yc015.github.io/reasoning-pr... cc @yidachen.bsky.social
Colorful depictions of reasoning progress: most of the time the system settles on the correct answer but sometimes it vacillates in interesting ways.
wattenberg.bsky.social
Very cool! You're definitely not alone in finding this fascinating. If you're looking for other people interested in this kind of thing, drop by the Arbor Project page, if you haven't already. github.com/ArborProject...
GitHub - ARBORproject/arborproject.github.io
Contribute to ARBORproject/arborproject.github.io development by creating an account on GitHub.
github.com
wattenberg.bsky.social
The wind map at hint.fm/wind/ has been running since 2012, relying on weather data from NOAA. We added a notice like this today. Thanks to @cambecc.bsky.social for the inspiration.
wattenberg.bsky.social
It's based on a data set of multiple-choice questions that have a known right answer, so this visualization only works when you have labeled ground truth. Definitely wouldn't shock me if those answers were labeled by grad students, though!
wattenberg.bsky.social
Great questions! Maybe it would be faster... or maybe it's doing something important under the hood that we can't see? I genuinely have no idea.
wattenberg.bsky.social
We also see cases where it starts out with the right answer, but eventually "convinces itself" of the wrong answer! I would love to understand the dynamics better.
wattenberg.bsky.social
Neat visualization that came up in the ARBOR project: this shows DeepSeek "thinking" about a question, and color is the probability that, if it exited thinking, it would give the right answer. (Here yellow means correct.)
wattenberg.bsky.social
Thank you! That's a great write-up, and this is definitely an interesting experiment. The distinction between how the model might do parsing vs. solving is very much worth thinking about. I added a few thoughts on the wiki page. github.com/ARBORproject...
Chain of Thought for Tsumego (Go Life or Death) Problems
Contribute to ARBORproject/arborproject.github.io development by creating an account on GitHub.
github.com
wattenberg.bsky.social
Great thread describing the new ARBOR open interpretability project, which has some fascinating projects already. Take a look!
ARBOR aims to accelerate the internal investigation of the new class of AI "reasoning" models.

See the ARBOR discussion board for a thread for each project underway.

github.com/ArborProjec...