Andreas Kirsch
@blackhc.bsky.social
4.3K followers 2K following 130 posts
My opinions only here. 👨‍🔬 RS DeepMind Past: 👨‍🔬 R Midjourney 1y 🧑‍🎓 DPhil AIMS Uni of Oxford 4.5y 🧙‍♂️ RE DeepMind 1y 📺 SWE Google 3y 🎓 TUM 👤 @nwspk
Posts Media Videos Starter Packs
Pinned
blackhc.bsky.social
Ever wondered why presenting more facts can sometimes *worsen* disagreements, even among rational people? 🤔

It turns out, Bayesian reasoning has some surprising answers - no cognitive biases needed! Let's explore this fascinating paradox quickly ☺️
blackhc.bsky.social
100000t of bombs vs the civilian casualty rates in Gaza actually shows that the IDF went out of their way to protect the civilian population

Compare this to, I dunno, the Dresden bomb raids:

25000 deaths over 4 days from just 3900t of bombs

Was that genocide?

en.m.wikipedia.org/wiki/Bombing...
Bombing of Dresden - Wikipedia
en.m.wikipedia.org
blackhc.bsky.social
Lol this fake statistician has blocked me for calling out his stupidity 😅
blackhc.bsky.social
Because the urban warfare during the counterinsurgency, 60% of the buildings were destroyed during fighting in the course of a couple of months. This is in line with the report you cited above.

How is this war in Gaza different to that war?

en.m.wikipedia.org/wiki/Falluja...
blackhc.bsky.social
Did the US also commit genocide in Iraq?
blackhc.bsky.social
The Gedankenexperiment says they must be doing a pretty terrible job if this is supposed to be a genocide 🙄
blackhc.bsky.social
x.com/BlackHC/stat...

Some thoughts on the comment paper
Reposted by Andreas Kirsch
lambda.bsky.social
We launched CoverDrop 🎉 providing sources with a secure and anonymous way to talk to journalists. Having started five years ago as a PhD research project, this now ships within the Guardian app to millions of users—all of which provide cover traffic. Paper, code, and more info: www.coverdrop.org
CoverDrop: Blowing the Whistle Through A News App
www.coverdrop.org
blackhc.bsky.social
Oh yeah I can't believe ai generated ASMR is also taking off. I've seen one or two of those!
blackhc.bsky.social
What's your favorite Veo video?
blackhc.bsky.social
Yeah but that's where most of the interesting things are 😅
blackhc.bsky.social
I hope the authors (I QT'ed @MFarajtabar above) can revisit the Tower of Hanoi results and examine the confounders to strengthen the paper (or just drop ToH). This will help keep the focus on the more interesting other environments for which the claims in the paper seem valid 🙏
blackhc.bsky.social
And other influencers exaggerate its results massively:

x.com/RubenHssd/s...
blackhc.bsky.social
So thinks are not as bleak as the coverage makes them sound. Aprospos coverage: here are some glowing reviews of the paper that do not question it:

Of course, @GaryMarcus likes it very much:

garymarcus.substack.com/p/a-knockou...
A knockout blow for LLMs?
LLM “reasoning” is so cooked they turned my name into a verb
garymarcus.substack.com
blackhc.bsky.social
I want to point to one more claim which is already outdated (the relevant paper was only published a few days ago so hardly anyone's fault):

The ProRL paper by Nvidia has shown that RL-based models can truly learn new things - if you run RL long enough!

arxiv.org/abs/2505.24864
blackhc.bsky.social
But as Gemini 2.5 Pro explains, River Crossing's optimal solutions are rather short, but they have a high branching factor and number of possible states with dead-ends.

That models fail here is a lot more interesting and points towards areas of improvements.
blackhc.bsky.social
All in all, the Tower of Hanoi results cannot be given any credence because it seems there are many confounders and simpler explanations.

However, I don't think the other games hit the same issues. If we look at River Crossing it seems to hit high token counts very quickly:
blackhc.bsky.social
Thus, @scaling01 calls out their conclusion: it looks like they didn't pay as close attention to the model's reasoning traces as they have claimed 😬

x.com/scaling01/s...
blackhc.bsky.social
This can also be observed in other simpler non-reasoning tasks:

x.com/Afinetheore...
blackhc.bsky.social
New LRMs are actually trained to be aware of their token limits. If they cannot think there way through, they find other solutions.

Apparently, the models say so at N=9 in ToH or refuse the write out the whole solution "by hand" and write code instead:

x.com/scaling01/s...
blackhc.bsky.social
Another "counterintuitive" behavior that is reported that the models start outputting fewer thinking tokens as the problems get more complex (e.g., more disks in Tower of Hanoi).

But again, this can be explained away sadly for ToH (but only ToH!).

x.com/MFarajtabar...
blackhc.bsky.social
Because, wait, it gets worse:

For N >= 12 or 13 disks, the LRM could not even output all the moves for the solution even if it wanted because the models can only output 64k tokens. Half of the plots of the ToH are essentially meaningless anyway.

x.com/scaling01/s...
blackhc.bsky.social
This is very important: we do not need the LRM to be bad at reasoning for this to happen.

It just happens because the correct solution is so long and is not allowed to contain any typos.

(IMO the paper should really consider dropping the ToH environment from their results.)
blackhc.bsky.social
It assumes the LLM samples the correct token with a given (high) probability and looks at the required number of tokens we need to output all the moves. Then:

p("all correct") = p**(2^N - 1)

and p=0.999 matches the ToH results in the paper above.
blackhc.bsky.social
We use top-p or top-k sampling with a temperature of 0.7 after all. Thus, there is a chance the model gets unlucky and the wrong token is sampled, resulting in failure. (They should really try min-p 😀)

@scaling01 has a nice toy model that matches the paper results: