Ahmad Beirami
@abeirami.bsky.social
3.6K followers 1.3K following 320 posts
something new // Gemini RL+inference @ Google DeepMind // Conversational AI @ Meta // RL Agents @ EA // ML+Information Theory @ MIT+Harvard+Duke // Georgia Tech PhD // زن زندگی آزادی 📍{NYC, SFO, YYZ} 🔗 https://beirami.github.io/
Posts Media Videos Starter Packs
Pinned
abeirami.bsky.social
After three incredible years, today is my last day at Google DeepMind!

I am truly grateful to the amazing colleagues who made the journey 1000x more fruitful and enjoyable! I am forever indebted to my collaborators who showed me how to be better at everything via demonstrations.
abeirami.bsky.social
The math that LLMs can do today is novel enough to be considered publishable, but it's not the kind of math that would be consequential.
abeirami.bsky.social
Let's regress from here to AGI!
nagpalchirag.bsky.social
Kernel regression is the only way to achieve AGI
abeirami.bsky.social
This is the conclusion slide of a talk I gave more than a year ago on RL/Alignment! It still holds true today.
Slide titled “Takeaways (alignment recipe).”

Step 1: Perform Best-of-n and make sure it works as desired.
– Inspect a few responses and verify the reward-induced ranking makes sense.
– Best-of-n gives the best trade-offs; if it doesn’t work, no fancy method will.
– You can debug best-of-n much faster.

Step 2: Only then train your favorite alignment method.
– Track KL(π‖p) throughout training:
• KL > 100: results are unlikely to be useful.
• KL > 15: inspect outcomes for reward hacking.
• KL < 8: you are probably OK.

Bottom banner in a black box repeats “(1) Look at your data! (2) Look at your data! (3) Look at your data!” in blue, green, and red.
abeirami.bsky.social
This also applies to telling your story (e.g., in a CV, bio, interview, etc).

Focus on what you have accomplished and what you are excited about doing next; not just where you did it!
abeirami.bsky.social
I occasionally get messages asking how to follow my path and get into Meta, DeepMind, or similar places. That is the wrong question. Do not focus on the brand! Focus on what you want to work on, then find the opportunity that fits your goals best.
abeirami.bsky.social
Haha. the content was:

If a paper is great, the credit goes to the first author.

If a paper has any flaws, the responsibility falls on the last author.
Reposted by Ahmad Beirami
shubhendu.bsky.social
The actual unpopular opinion is that the notion of senior and junior authors should be abolished. It has completely diluted the notion of scientific authorship and created this entire industry of free-riding, head-in-the-clouds, incompetent PIs/managers. List down exact contributions instead. [+]
abeirami.bsky.social
Unpopular opinion:
When a paper has a senior mentor and a junior mentee, the senior author must make sure the claims are correct and well supported. They must check every claim and gate the submission until it meets that bar.
abeirami.bsky.social
Glad you asked.
bsky.app/profile/abei...
abeirami.bsky.social
Corollary: If you lack bandwidth or expertise to act as the verifier, then you shouldn't sign up to be the senior author of a paper!
abeirami.bsky.social
Unpopular opinion:
When a paper has a senior mentor and a junior mentee, the senior author must make sure the claims are correct and well supported. They must check every claim and gate the submission until it meets that bar.
abeirami.bsky.social
I occasionally get messages asking how to follow my path and get into Meta, DeepMind, or similar places. That is the wrong question. Do not focus on the brand! Focus on what you want to work on, then find the opportunity that fits your goals best.
Reposted by Ahmad Beirami
eugenevinitsky.bsky.social
Related to this if a paper turns out to have a major error in it, you’re supposed to throw yourself under the bus not your students.
abeirami.bsky.social
Unpopular opinion:
When a paper has a senior mentor and a junior mentee, the senior author must make sure the claims are correct and well supported. They must check every claim and gate the submission until it meets that bar.
abeirami.bsky.social
I think for every deliverable, there has to be one person who is responsible (gets it done) and one person who is accountable (makes sure it's done correctly).

Middle authors can be responsible or accountable for a subset of tasks.
abeirami.bsky.social
Corollary: If you lack bandwidth or expertise to act as the verifier, then you shouldn't sign up to be the senior author of a paper!
abeirami.bsky.social
Unpopular opinion:
When a paper has a senior mentor and a junior mentee, the senior author must make sure the claims are correct and well supported. They must check every claim and gate the submission until it meets that bar.
abeirami.bsky.social
The junior author is the generator. The senior author is the verifier. The verifier should teach/distill some checks to the generator, but the verifier keeps final responsibility. If a wrong claim gets out, it is on the verifier!
abeirami.bsky.social
Unpopular opinion:
When a paper has a senior mentor and a junior mentee, the senior author must make sure the claims are correct and well supported. They must check every claim and gate the submission until it meets that bar.
abeirami.bsky.social
This is the recipe for many provable claims:

Make enough assumptions and narrow down the claim, then prove a narrow result with caveats. Present it as broad, hide the caveats, and declare “XYZ is provable!”
abeirami.bsky.social
Today, Industry research is focused on short term (3-6months) bets. Academics have an opportunity to balance their portfolio with medium term (1-2 years) and long term (5-10 years) bets. Putting all academic efforts in short-term basket is suboptimal!
abeirami.bsky.social
When I worked in corporate, I was often first in the office because that routine worked for me. It was a personal preference, not a benchmark for anyone else.

We should not judge commitment by hours, especially in research. We should look for thoughtful work and steady progress.
abeirami.bsky.social
True, this sounds obvious but it is more common than we'd hope, unfortunately.
abeirami.bsky.social
Common mistake in LLM prompting projects: jumping into full-scale pipelines (datasets and inference) without testing feasibility. Iterating at scale is expensive and time-consuming.

Start with ONE example to validate the hypothesis, verify context, debug the design, then scale.
abeirami.bsky.social
I have to admit that I embarrassingly didn't know about this history. Nice reading material for the long weekend, thank you :)
abeirami.bsky.social
But I don't disagree that maybe it is already better than an average human, because we humans are pretty bad at this too.
abeirami.bsky.social
And audience modeling is a hard problem and won't be solved unless we effectively solve memory.

On the clarity point: I think we are still not there yet on info seeking queries, e.g., GPT5 Thinking returns a lot of info (high recall) but is unable to put the info in a clear and logical structure.
abeirami.bsky.social
The open question is how to choose annotators who can teach that skill. We need people who understand the point in the first place, and who can tell when a point is vague, what context is missing, and how to fix it.