Yonatan Belinkov ✈️ COLM2025
@boknilev.bsky.social
110 followers 390 following 19 posts
Assistant professor of computer science at Technion; visiting scholar at @KempnerInst 2025-2026 https://belinkov.com/
Posts Media Videos Starter Packs
Reposted by Yonatan Belinkov ✈️ COLM2025
mtutek.bsky.social
🤔What happens when LLM agents choose between achieving their goals and avoiding harm to humans in realistic management scenarios? Are LLMs pragmatic or prefer to avoid human harm?

🚀 New paper out: ManagerBench: Evaluating the Safety-Pragmatism Trade-off in Autonomous LLMs🚀🧵
boknilev.bsky.social
Reach out if you'd like to discuss anything related to model interpretability and controllability, robustness, multi-agent communication, biological LMs, etc.
Also happy to talk about PhD and Post-doc opportunities!
boknilev.bsky.social
In #Interplay25 workshop, Friday ~11:30, I'll present on measuring *parametric* CoT faithfulness on behalf of
@mtutek.bsky.social , who couldn't travel:
bsky.app/profile/mtut...

Later that day we'll have a poster on predicting success of model editing by Yanay Soker, who also couldn't travel
boknilev.bsky.social
@zorikgekhman.bsky.social
will present on Wednesday a poster on hidden factual knowledge in LMs
bsky.app/profile/zori...
zorikgekhman.bsky.social
🚨 It's often claimed that LLMs know more facts than they show in their outputs, but what does this actually mean, and how can we measure this “hidden knowledge”?

In our new paper, we clearly define this concept and design controlled experiments to test it.
1/🧵
boknilev.bsky.social
@itay-itzhak.bsky.social
presenting today, morning, a spotlight talk and poster on the origin of cognitive biases in LLMs
bsky.app/profile/itay...
itay-itzhak.bsky.social
🚨New paper alert🚨

🧠
Instruction-tuned LLMs show amplified cognitive biases — but are these new behaviors, or pretraining ghosts resurfacing?

Excited to share our new paper, accepted to CoLM 2025🎉!
See thread below 👇
#BiasInAI #LLMs #MachineLearning #NLProc
boknilev.bsky.social
Traveling to #COLM2025 this week, and here's some work from our group and collaborators:
Cognitive biases, hidden knowledge, CoT faithfulness, model editing, and LM4Science
See the thread for details and reach out if you'd like to discuss more!
Reposted by Yonatan Belinkov ✈️ COLM2025
amuuueller.bsky.social
What's the right unit of analysis for understanding LLM internals? We explore in our mech interp survey (a major update from our 2024 ms).

We’ve added more recent work and more immediately actionable directions for future work. Now published in Computational Linguistics!
boknilev.bsky.social
Opportunities to join my group in fall 2026:
* PhD applications direct or via ELLIS @ellis.eu (ellis.eu/news/ellis-p...)
* Post-doc applications direct or via Azrieli (azrielifoundation.org/fellows/inte...) or Zuckerman (zuckermanstem.org/ourprograms/...)
boknilev.bsky.social
Excited to join @KempnerInst this year!
Get in touch if you're in the Boston area and want to chat about anything related to AI interpretability, robustness, interventions, safety, multi-modality, protein/DNA LMs, new architectures, multi-agent communication, or anything else you're excited about!
Reposted by Yonatan Belinkov ✈️ COLM2025
mtutek.bsky.social
Thrilled that FUR was accepted to @emnlpmeeting.bsky.social Main🎉

In case you can’t wait so long to hear about it in person, it will also be presented as an oral at @interplay-workshop.bsky.social @colmweb.org 🥳

FUR is a parametric test assessing whether CoTs faithfully verbalize latent reasoning.
boknilev.bsky.social
BlackboxNLP is the workshop on interpreting and analyzing NLP models (including LLMs, VLMs, etc). We accept full (archival) papers and extended abstracts.

The workshop is highly attended and is a great exposure for your finished work or feedback on work in progress.

#emnlp2025 at Sujhou, China!
blackboxnlp.bsky.social
📢 Call for Papers! 📢
#BlackboxNLP 2025 invites the submission of archival and non-archival papers on interpreting and explaining NLP models.

📅 Deadlines: Aug 15 (direct submissions), Sept 5 (ARR commitment)
🔗 More details: blackboxnlp.github.io/2025/call/
boknilev.bsky.social
Join our Discord for discussions and a bunch of simple submission ideas you can try!
discord.gg/n5uwjQcxPR

Participants will have the option to write a system description paper that gets published.
Reposted by Yonatan Belinkov ✈️ COLM2025
blackboxnlp.bsky.social
Have you started working on your submission for the MIB shared task yet? Tell us what you’re exploring!

New featurization methods?
Circuit pruning?
Better feature attribution?

We'd love to hear about it 👇
Reposted by Yonatan Belinkov ✈️ COLM2025
blackboxnlp.bsky.social
Working on feature attribution, circuit discovery, feature alignment, or sparse coding?
Consider submitting your work to the MIB Shared Task, part of this year’s #BlackboxNLP

We welcome submissions of both existing methods and new or experimental POCs!
Reposted by Yonatan Belinkov ✈️ COLM2025
danaarad.bsky.social
VLMs perform better on questions about text than when answering the same questions about images - but why? and how can we fix it?

In a new project led by Yaniv (@YNikankin on the other app), we investigate this gap from an mechanistic perspective, and use our findings to close a third of it! 🧵
Reposted by Yonatan Belinkov ✈️ COLM2025
danaarad.bsky.social
Tried steering with SAEs and found that not all features behave as expected?

Check out our new preprint - "SAEs Are Good for Steering - If You Select the Right Features" 🧵
Reposted by Yonatan Belinkov ✈️ COLM2025
tomerashuach.bsky.social
🚨New paper at #ACL2025 Findings!
REVS: Unlearning Sensitive Information in LMs via Rank Editing in the Vocabulary Space.
LMs memorize and leak sensitive data—emails, SSNs, URLs from their training.
We propose a surgical method to unlearn it.
🧵👇w/ @boknilev.bsky.social @mtutek.bsky.social
1/8
boknilev.bsky.social
Interested in mechanistic interpretability and care about evaluation? Please consider submitting to our shared task at #blackboxNLP this year!
blackboxnlp.bsky.social
BlackboxNLP, the leading workshop on interpretability and analysis of language models, will be co-located with EMNLP 2025 in Suzhou this November! 📆

This edition will feature a new shared task on circuits/causal variable localization in LMs, details here: blackboxnlp.github.io/2025/task
Reposted by Yonatan Belinkov ✈️ COLM2025
boknilev.bsky.social
I think a scalable open source implementation would have many uses! Let’s say I can’t run all pretraining data because of cost. And I run a subset and get influential examples. What would that mean concerning what I’m missing?
boknilev.bsky.social
Looks great! What would it take to run this on another model and dataset?
boknilev.bsky.social
This has been a huge team effort with many talented contributors. Very thankful for everyone’s contributions!

See the list here:
bsky.app/profile/amuu...
amuuueller.bsky.social
This was a huge collaboration with many great folks! If you get a chance, be sure to talk to Atticus Geiger, @sarah-nlp.bsky.social, @danaarad.bsky.social, Iván Arcuschin, @adambelfki.bsky.social, @yiksiu.bsky.social, Jaden Fiotto-Kaufmann, @talhaklay.bsky.social, @michaelwhanna.bsky.social, ...