micha heilbron
@mheilbron.bsky.social
780 followers 320 following 64 posts
Assistant Professor of Cognitive AI @UvA Amsterdam language and vision in brains & machines cognitive science 🤝 AI 🤝 cognitive neuroscience michaheilbron.github.io
Posts Media Videos Starter Packs
mheilbron.bsky.social
omg. what journal? name and shame
mheilbron.bsky.social
huh! if these effects are similar and consistent, I think it should work, but the q. is how do you get a vector representation for novel pseudowords? we currently use lexicosemantic word vectors and they are undefined for novel words.

so how to represent the novel words? v. interesting test case
mheilbron.bsky.social
Together, our results support a classic idea: cognitive limitations can be a powerful inductive bias for learning

Yet they also reveal a curious distinction: a model with more human-like *constraints* is not necessarily more human-like in its predictions
mheilbron.bsky.social
This paradox – better language models yielding worse behavioural predictions – could not be explained by prior explanations: The mechanism appears distinct from those linked to superhuman training scale or memorisation
mheilbron.bsky.social
However, we then used these models to predict human behaviour

Strikingly these same models that were demonstrably better at the language task, were worse at predicting human reading behaviour
mheilbron.bsky.social
The benefit was robust

Fleeting memory models achieved better next-token prediction (lower loss) and better syntactic knowledge (higher accuracy) on the BLiMP benchmark

This was consistent across seeds and for both 10M and 100M training sets
mheilbron.bsky.social
But we noticed this naive decay was too strong

Human memory has a brief 'echoic' buffer that perfectly preserves the immediate past. When we added this – a short window of perfect retention before the decay -- the pattern flipped

Now, fleeting memory *helped* (lower loss)
mheilbron.bsky.social
Our first attempt, a "naive" memory decay starting from the most recent word, actually *impaired* language learning. Models with this decay had higher validation loss, and this worsened (even higher loss) as the decay became stronger
mheilbron.bsky.social
To test this in a modern context, we propose the ‘fleeting memory transformer’

We applied a power-law memory decay to the self-attention scores, simulating how access to past words fades over time, and ran controlled experiments on the developmentally realistic BabyLM corpus
mheilbron.bsky.social
However, this appears difficult to reconcile with the success of transformers, which can learn language very effectively, despite lacking working memory limitations or other recency biases

Would the blessing of fleeting memory still hold in transformer language models?
mheilbron.bsky.social
A core idea in cognitive science is that the fleetingness of working memory isn't a flaw

It may actually help at learning language by forcing a focus on the recent past and providing an incentive to discover abstract structure rather than surface details
mheilbron.bsky.social
New preprint! w/@drhanjones.bsky.social

Adding human-like memory limitations to transformers improves language learning, but impairs reading time prediction

This supports ideas from cognitive science but complicates the link between architecture and behavioural prediction
arxiv.org/abs/2508.05803
Human-like fleeting memory improves language learning but impairs reading time prediction in transformer language models
Human memory is fleeting. As words are processed, the exact wordforms that make up incoming sentences are rapidly lost. Cognitive scientists have long believed that this limitation of memory may, para...
arxiv.org
mheilbron.bsky.social
On Wednesday, Maithe van Noort will present a poster on “Compositional Meaning in Vision-Language Models and the Brain”

First results from a much larger project on visual and linguistic meaning in brains and machines, with many collaborators -- more to come! 

t.ly/TWsyT
Poster Presentation
t.ly
mheilbron.bsky.social
On Friday, during a contributed talk (and a poster), @wiegerscheurer will present the project he spearheaded: “A hierarchy of spatial predictions across human visual cortex during natural vision” 

(Full preprint soon)

t.ly/fTJqy
Poster Presentation
t.ly
mheilbron.bsky.social
CCN has arrived here here in Amsterdam!

Come find me to meet or catch up

Some highlights from students and collaborators:
mheilbron.bsky.social
Waarom vergeet je namen maar weet je nog precies wat iemand doet? En zijn herinneringen ooit echt helemaal weg?

Ik ging bij Oplossing Gezocht in gesprek over hoe ons brein informatie opslaat en waarom vergeten eigenlijk heel slim is:
www.nemokennislink.nl/publicaties/...
Waarom kom ik toch niet op die naam?
Eh, je weet wel wie... dinges! Heb jij ook soms zo’n moeite om op een naam te komen? Hersenonderzoeker Micha Heilbron legt uit hoe dat komt - en waarom een naam eigenlijk niet zo belangrijk is.
www.nemokennislink.nl
Reposted by micha heilbron
timkietzmann.bsky.social
Exciting new preprint from the lab: “Adopting a human developmental visual diet yields robust, shape-based AI vision”. A most wonderful case where brain inspiration massively improved AI solutions.

Work with @zejinlu.bsky.social @sushrutthorat.bsky.social and Radek Cichy

arxiv.org/abs/2507.03168
arxiv.org
mheilbron.bsky.social
i’m all in the “this is a neat way to help explain things” camp fwiw :)
mheilbron.bsky.social
Our findings, together with some other recent studies, suggest the brain may use a similar strategy — constantly predicting higher-level features — to efficiently learn robust visual representations of (and from!) the natural world
mheilbron.bsky.social
This preference for higher-level information departs from traditional predictive coding -- but aligns with recent, successful algorithms in AI for predictive self-supervised learning, which encourage predicting higher rather than lower-level visual features (e.g. MAE, CPC, JEPA)
mheilbron.bsky.social
So, what does this all mean?

The visual system seems to be constantly engaged in a sophisticated guessing game, predicting sensory input based on context

But interestingly, it seems to predict more abstract, higher-level properties, even in the earliest stages of cortex