Naomi Saphra
@nsaphra.bsky.social
9.7K followers 1.5K following 2.4K posts
Waiting on a robot body. All opinions are universal and held by both employers and family. Literally a professor. Recruiting students to start my lab. ML/NLP/they/she.
Posts Media Videos Starter Packs
Pinned
nsaphra.bsky.social
I wrote something up for AI people who want to get into bluesky and either couldn't assemble an exciting feed or gave up doomscrolling when their Following feed switched to talking politics 24/7.
The AI Researcher's Guide to a Non-Boring Bluesky Feed | Naomi Saphra
How to migrate to bsky without a boring feed.
nsaphra.net
Reposted by Naomi Saphra
vgel.me
if you're interesting in gaining a better intuition for how llms behave at inference time, you should try logitloom🌱, the open-source tool i made for exploring token trajectory trees (aka looming) on base and instruct models! more info in thread

🌱 vgel.me/logitloom
💻 github.com/vgel/logitloom
nsaphra.bsky.social
How many flagged images could you look at before you lose the ability to look normally at a child? At a puppy? At any adult woman? At any black person? I would rather give up visual social media entirely than increase the number of human casualties.
nsaphra.bsky.social
It is wild that anyone out there is advocating for MORE human labor in image moderation. a friend of mine used to be a 4chan janitor. they’re dead now, but they spent the rest of their life unable to look at a child without feeling sick.
nsaphra.bsky.social
Yeah. The huggingface incident was about someone collecting a small, public, open access dataset to support others’ data analysis.
Every company whose product is an actual model (and not scientific infrastructure) has quietly scraped far more.
nsaphra.bsky.social
They want more traumatized human moderators forced to flip through reams of CP and snuff tyvm. Can’t have clankers steal these key human jobs which have a turnover rate of well under a year because it only takes a couple weeks to get intractable PTSD
nsaphra.bsky.social
One-in-a-million events happen all the time and many will happen to you eventually. Live like an actuary.
nsaphra.bsky.social
It is very grounding to calculate how many people in your country are dealing with a newly diagnosed physical disability, a child with cancer, a parent with dementia, the death of a spouse. Nothing will make you feel smaller than to look up how common the worst thing that’s ever happened to you is.
Reposted by Naomi Saphra
dorialexander.bsky.social
Ok while I do think many issues with AI are overblown, web agents hitting hard on infra sustainability and knowledge access. Was just browsing the reference database for Disney comics: now login only indefinitely.
Reposted by Naomi Saphra
gregdnlp.bsky.social
Find my students and collaborators at COLM this week!

Tuesday morning: @juand-r.bsky.social and @ramyanamuduri.bsky.social 's papers (find them if you missed it!)

Wednesday pm: @manyawadhwa.bsky.social 's EvalAgent

Thursday am: @anirudhkhatry.bsky.social 's CRUST-Bench oral spotlight + poster
Reposted by Naomi Saphra
aaup.org
AAUP @aaup.org · 18h
NO LOYALTY OATHS IN HIGHER ED!

Trump's attacks on our universities are an attempt to consolidate power. This loyalty oath directly undermines our right to academic freedom & goes against every democratic principle our country should uphold.

Please click below, sign, & share.

#DefendHigherEd
University Administrations: Reject Trump's "Loyalty Oath" Compacts
The Trump administration is trying to blackmail schools to let him and his unqualified bureaucrats run our schools. They want to dictate what schools teach, who they admit and hire, what researchers s...
actionnetwork.org
nsaphra.bsky.social
Omg congrats @ajyl.bsky.social
colmweb.org
Outstanding paper 2🏆: Shared Global and Local Geometry of Language Model Embeddings
openreview.net/forum?id=aJD...
Reposted by Naomi Saphra
jennhu.bsky.social
At #COLM2025 and would love to chat all things cogsci, LMs, & interpretability 🍁🥯 I'm also recruiting!

👉 I'm presenting at two workshops (PragLM, Visions) on Fri

👉 Also check out "Language Models Fail to Introspect About Their Knowledge of Language" (presented by @siyuansong.bsky.social Tue 11-1)
Reposted by Naomi Saphra
mariaa.bsky.social
Here’s a #COLM2025 feed!

Pin it 📌 to follow along with the conference this week!
nsaphra.bsky.social
It’s a drawing my sister made for my bat mitzvah when I was 13, because I liked octopuses!
Reposted by Naomi Saphra
quantamagazine.bsky.social
An unexpected challenge at the start of Naomi Saphra’s career shaped her research as a computer scientist. www.quantamagazine.org/to-understan...
Reposted by Naomi Saphra
vgel.me
new blog post! why do LLMs freak out over the seahorse emoji? i put llama-3.3-70b through its paces with the logit lens to find out, and explain what the logit lens (everyone's favorite underrated interpretability tool) is in the process.

link in reply!
nsaphra.bsky.social
do you think it’s ok to skip your 1500th wedding anniversary??? this is a special moment. congratulations to all bats
Reposted by Naomi Saphra
gretatuckute.bsky.social
Check out @mryskina.bsky.social's talk and poster at COLM on Tuesday—we present a method to identify 'semantically consistent' brain regions (responding to concepts across modalities) and show that more semantically consistent brain regions are better predicted by LLMs.
mryskina.bsky.social
Interested in language models, brains, and concepts? Check out our COLM 2025 🔦 Spotlight paper!

(And if you’re at COLM, come hear about it on Tuesday – sessions Spotlight 2 & Poster 2)!
Paper title: Language models align with brain regions that represent concepts across modalities.
Authors:  Maria Ryskina, Greta Tuckute, Alexander Fung, Ashley Malkin, Evelina Fedorenko. 
Affiliations: Maria is affiliated with the Vector Institute for AI, but the work was done at MIT. All other authors are affiliated with MIT. 
Email address: maria.ryskina@vectorinstitute.ai.
nsaphra.bsky.social
COLM is approaching! You should meet our lead author Natalie and ask her about how LMs learn to beat human experts!
nsaphra.bsky.social
How can an imitative model like an LLM outperform the experts it is trained on? Our new COLM paper outlines three types of transcendence and shows that each one relies on a different aspect of data diversity. arxiv.org/abs/2508.17669
nsaphra.bsky.social
I think I’ve seen work on clustering ensembles and pruning out underperforming members of each cluster but not sure if it works in modern scales
nsaphra.bsky.social
My friends and family know I love looking at random variation. But my ENEMIES claim random seed will be UNIMPORTANT at large scales. Fools! I will say YOU ARE WRONG, MY ENEMIES. LOOK.
nsaphra.bsky.social
Ensembling is good in data bottlenecked regimes! My takeaway: random variation is only going to get more important at scale—even as it decreases in magnitude. The deterministic benefits of scale eventually decline faster than a model’s randomly selected advantages!
Pre-training under infinite compute
Since compute grows much faster than web text available for language model pre-training, we ask how one should approach pre-training under fixed data and no compute constraints. We first show that exi...
arxiv.org
nsaphra.bsky.social
A preliminary result that ChatGPT-3.5 was more likely to answer an under specified request directly, rather than ask necessary follow up questions, if you introduced yourself with a female name. We were gonna call it The Mansplaining Effect. Wound up focusing on guardrails in the paper instead!
nsaphra.bsky.social
I have posted stuff on here that I probably shouldn’t have, so imagine what I post to the groupchat