Martin Hebart
@martinhebart.bsky.social
4.6K followers 550 following 320 posts
Proud dad, Professor of Computational Cognitive Neuroscience, author of The Decoding Toolbox, founder of http://things-initiative.org our lab 👉 https://hebartlab.com
Posts Media Videos Starter Packs
Reposted by Martin Hebart
patxelos.bsky.social
1/Preprint Alert🔔: Across two experiments plus a computational model, we show the visual system compresses complex scenes into summary statistics that can guide behavior without conscious access to the task-defining features. We term this the Ensemble Blindsight effect.
Reposted by Martin Hebart
malcolmgcampbell.bsky.social
🚨Our preprint is online!🚨

www.biorxiv.org/content/10.1...

How do #dopamine neurons perform the key calculations in reinforcement #learning?

Read on to find out more! 🧵
martinhebart.bsky.social
I’m also curious to see the manuscript. I’d like to understand better how this is smarter than hierarchical partitioning (i.e. non-Bayesian model averaging ;) since this is just a different way to assign the shared variance to unique parts, but it doesn’t solve the indeterminacy. Can BMA address it?
martinhebart.bsky.social
Can’t we compute an upper bound for the “shared variance”, i.e. the area of indistinguishability, by simply taking 1 minus the sum of the unique parts of all other variables, since the shared cannot possibly explain more than that?

And a lower by assuming all other variables are unrelated to Y?
Reposted by Martin Hebart
joachimbaumann.bsky.social
🚨 New paper alert 🚨 Using LLMs as data annotators, you can produce any scientific result you want. We call this **LLM Hacking**.

Paper: arxiv.org/pdf/2509.08825
We present our new preprint titled "Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation".
We quantify LLM hacking risk through systematic replication of 37 diverse computational social science annotation tasks.
For these tasks, we use a combined set of 2,361 realistic hypotheses that researchers might test using these annotations.
Then, we collect 13 million LLM annotations across plausible LLM configurations.
These annotations feed into 1.4 million regressions testing the hypotheses. 
For a hypothesis with no true effect (ground truth $p > 0.05$), different LLM configurations yield conflicting conclusions.
Checkmarks indicate correct statistical conclusions matching ground truth; crosses indicate LLM hacking -- incorrect conclusions due to annotation errors.
Across all experiments, LLM hacking occurs in 31-50\% of cases even with highly capable models.
Since minor configuration changes can flip scientific conclusions, from correct to incorrect, LLM hacking can be exploited to present anything as statistically significant.
martinhebart.bsky.social
P.P.S.: re-reading this, I notice the inaccuracies in my language (e.g. not expected by chance - expected given no actually shared variance). Apologies! I just quickly typed my thoughts down, hopefully in an understandable fashion!
martinhebart.bsky.social
P.S.: you might wonder where bounds for shared variance could be helpful. If your model doesn’t explain more shared variance than expected by the baseline (which is given by the lower bound), you can confidently say there is no evidence for shared variance.
martinhebart.bsky.social
6. We now use hierarchical partitioning to address interpretability issues for “unique parts” in the presence of highly collinear variables. I saw @diedrichsenjorn.bsky.social has a Bayesian method up his sleeve that is a variant of hierarchical partitioning, so it’s worth keeping an eye out for it.
martinhebart.bsky.social
5. So while the interpretation of variance partitioning can be tricky w/ collinear variables, bounds can at least help limit the range of possibilities. Curious if my intuition is correct or if anyone has looked into this.
martinhebart.bsky.social
4. Unique variance becomes very hard to interpret when we have more than 2 independent variables if two of them are highly collinear, because it may seem as if the third variable is better. Again, the upper bounds are really relevant here.
martinhebart.bsky.social
3. However, we can compute bounds for the shared variance. I wonder if this has ever been done. Assuming correlations of x1 with x2, x1 with y and setting r(x2,y)=0, we can figure out what shared variance we expect by chance. This doesn’t solve the interpretation problem but it constrains it.
martinhebart.bsky.social
2. The shared variance just means the part that cannot be assigned uniquely, and there are different ways of fairly distributing variance but we cannot figure out where the variance is coming from.
martinhebart.bsky.social
I wanted to add some thoughts to this excellent blog post, not detailed, maybe wrong, maybe useful:
1. Unique variance is easy to interpret as a lower bound of what a variable explains (the upper bound being either what the variable explains alone or what the other variables cannot explain uniquely)
diedrichsenjorn.bsky.social
Variance partitioning is used to quantify the overlap of two models. Over the years, I have found that this can be a very confusing and misleading concept. So we finally we decided to write a short blog to explain why.
@martinhebart.bsky.social @gallantlab.org
diedrichsenlab.org/BrainDataSci...
Reposted by Martin Hebart
neurosteven.bsky.social
🧠 New preprint: Why do deep neural networks predict brain responses so well?
We find a striking dissociation: it’s not shared object recognition. Alignment is driven by sensitivity to texture-like local statistics.
📊 Study: n=57, 624k trials, 5 models doi.org/10.1101/2025...
Reposted by Martin Hebart
martinhebart.bsky.social
Pretty amazing how you can learn something in books and in school and it explains everything around you - how the planets move, etc. But then it’s just next level when you can see it is real, with your own eyes.
martinhebart.bsky.social
Cool stuff! So you are looking at the unique multivariate corr structure shared between two sets of variables, which provides a lower bound of the true unique correl.? Is there a difference between keeping one set of variables uncorrelated with the confounds vs. both? (akin to semi vs partial corr)
Reposted by Martin Hebart
s-michelmann.bsky.social
🚀Excited to share our project: Canonical Representational Mapping for Cognitive Neuroscience. @schottdorflab.bsky.social and I propose a novel multivariate method to isolate neural representations aligned with specific cognitive hypotheses 🧵https://www.biorxiv.org/content/10.1101/2025.09.01.673485v1
Reposted by Martin Hebart
imagingneurosci.bsky.social
Launched in 2023, Imaging Neuroscience is now firmly established, with full indexing (PubMed, etc.) and 700 papers to date.

We're very happy to announce that we are able to reduce the APC to $1400.

Huge thanks to all authors, reviewers, editorial team+board, and MIT Press.
martinhebart.bsky.social
It’s also really hard for me to see it as something different than a bat. And it’s also nonsense. It makes it even more interesting that kiddo saw it for longer periods of time in either state. Even though our brain can have competing interpretations, here the other interpretation doesn’t win over.
martinhebart.bsky.social
martinhebart.bsky.social
It’s easier to see if you focus on the bottom or top! It’s just a pattern of blobs. Do you see it now? On the left like bush or like half butterflies. On the right like eye glasses.
martinhebart.bsky.social
It’s easier to see if you focus on the bottom or top! It’s just a pattern of blobs. Do you see it now? On the left like bush or like half butterflies. On the right like eye glasses.
martinhebart.bsky.social
The photo was blurry, so that reduces the effect! It’s easier to see here.

I would also like to know more about what is known in that age range!