Charlotte Volk
@charlottevolk.bsky.social
260 followers 380 following 19 posts
MSc Student in NeuroAI @ McGill & Mila w/ Blake Richards & Shahab Bakhtiari
Posts Media Videos Starter Packs
Pinned
charlottevolk.bsky.social
🚨 New preprint alert!

🧠🤖
We propose a theory of how learning curriculum affects generalization through neural population dimensionality. Learning curriculum is a determining factor of neural dimensionality - where you start from determines where you end up.
🧠📈

A 🧵:

tinyurl.com/yr8tawj3
The curriculum effect in visual learning: the role of readout dimensionality
Generalization of visual perceptual learning (VPL) to unseen conditions varies across tasks. Previous work suggests that training curriculum may be integral to generalization, yet a theoretical explan...
tinyurl.com
charlottevolk.bsky.social
16. Second, what is considered hard for one person may not be as hard for another person due to past experiences, innate differences, etc. Following our theoretical rule to reach low-dimensional readout subspaces post-training calls for an individualized approach to curriculum design.
charlottevolk.bsky.social
15. First, easy and hard are not easily definable for every task. It worked well for our simple orientation discrimination task, but as tasks become more naturalistic and complex, defining easiness will not be straightforward. Neural data may be helpful for giving an objective measure of difficulty.
charlottevolk.bsky.social
14. In short:

Easy-to-hard learning curriculum (explicit or implicit) sets the dimensionality of the neural population recruited to solve the task + lower-d readout leads to better generalization.

But, there are some subtleties for applying this rule to the real world training design: 👇
charlottevolk.bsky.social
13. Is this low-d subspace what truly drives generalization? We tested this by training a model non-sequentially while transplanting the low-dimensional readout subspace from a different high-generalization model. We found that this partially "frozen" model could in fact generalize much better!
charlottevolk.bsky.social
12.

2) Initial training phase sets this dimensionality (measured with the Jaccard index). J = 1 → no change in the readout subspace

Therefore, learners following an explicit (or implicit) easy-to-hard curriculum will discover a lower-d readout subspace.
charlottevolk.bsky.social
11. But how does curriculum affect readout dimensionality?

Two steps:

1) Easy tasks lead to a lower-d readout subspace: larger angle separation → lower-d readout
charlottevolk.bsky.social
10. We measured the dimensionality of the models’ “readout subspace” - essentially, the dimensionality of the neural population that contributes most strongly to the model output. We found that the effective rank of the readout subspace directly correlates to transfer accuracy (i.e. generalization).
charlottevolk.bsky.social
9. We hypothesized that the efficacy of the learning curricula depends on how many distinct, useful visual features the brain recruits to solve the task - curricula which lead learners to rely on fewer, more essential visual features will result in better generalization.
charlottevolk.bsky.social
8. Interestingly, even in the shuffled curriculum, both humans and ANNs generalize better to new contexts when they focus on easy trials first, as measured by a “curriculum metric” in humans and the ratio of easy-to-hard samples used in the initial phase of shuffled training for the models.
charlottevolk.bsky.social
7. We found:
- Sequential and shuffled curricula significantly outperform a non-sequential baseline in ANNs & humans.
- Models do better on a sequential curriculum; human observers show comparable improvement on both sequential & shuffled, but with substantial variability in the shuffled curriculum.
charlottevolk.bsky.social
6. We trained humans and ANNs on orientation discrimination comparing 3 curricula:
1) A sequential easy-to-hard curriculum
2) A shuffled curriculum with randomly interleaved easy & hard trials
3) A non-sequential baseline with only hard trials.
We tested generalization on a hard transfer condition.
charlottevolk.bsky.social
5. In this study, we leveraged ANNs to develop a mechanistic predictive theory of learning generalization in humans. Specifically, we wanted to understand the role of **learning curriculum**, and develop a theory of how curriculum affects generalization.
charlottevolk.bsky.social
4. We know that artificial neural networks (ANNs) fail to generalize in similar ways to humans in simple visual learning tasks, thanks to the previous work of Wenliang and Seitz (2018) → more difficult training tasks lead to worse generalization, which is a phenomenon observed in humans and ANNs.
charlottevolk.bsky.social
3. But - people don’t *always* fail to generalize. Generalization is quite variable across tasks (Ahissar & Hochstein, 1997), and the reasons behind it are unclear. Hence, the importance of a theory of generalization → If you design a new training paradigm, you want to predict its generalization.
charlottevolk.bsky.social
2. Improving on simple visual tasks (e.g., texture discrimination) through practice does not necessarily transfer to a slightly different version of the same task (a new location or rotation). This has been known since the early 90s. (e.g. Karni and Sagi, 1991).
charlottevolk.bsky.social
1. Learning generalization has been one of the main focuses of any training domain, e.g., expert training, athletics, and rehabilitation. When you learn or improve on a skill, you want your improved skills to be applicable to new situations. But, humans don’t always generalize well to new contexts.
charlottevolk.bsky.social
🚨 New preprint alert!

🧠🤖
We propose a theory of how learning curriculum affects generalization through neural population dimensionality. Learning curriculum is a determining factor of neural dimensionality - where you start from determines where you end up.
🧠📈

A 🧵:

tinyurl.com/yr8tawj3
The curriculum effect in visual learning: the role of readout dimensionality
Generalization of visual perceptual learning (VPL) to unseen conditions varies across tasks. Previous work suggests that training curriculum may be integral to generalization, yet a theoretical explan...
tinyurl.com
Reposted by Charlotte Volk
hafezghm.bsky.social
Excited to share that seq-JEPA has been accepted to NeurIPS 2025!
hafezghm.bsky.social
Preprint Alert 🚀

Can we simultaneously learn transformation-invariant and transformation-equivariant representations with self-supervised learning?

TL;DR Yes! This is possible via simple predictive learning & architectural inductive biases – without extra loss terms and predictors!

🧵 (1/10)
Reposted by Charlotte Volk
averyryoo.bsky.social
New preprint! 🧠🤖

How do we build neural decoders that are:
⚡️ fast enough for real-time use
🎯 accurate across diverse tasks
🌍 generalizable to new sessions, subjects, and even species?

We present POSSM, a hybrid SSM architecture that optimizes for all three of these axes!

🧵1/7
charlottevolk.bsky.social
Excited to be at #Cosyne2025 for the first time! I'll be presenting my poster [2-104] during the Friday session. E-poster here: www.world-wide.org/cosyne-25/se...
Reposted by Charlotte Volk
shahabbakht.bsky.social
📢 We have a new #NeuroAI postdoctoral position in the lab!

If you have a strong background in #NeuroAI or computational neuroscience, I’d love to hear from you.

(Repost please)

🧠📈🤖