Andrea de Varda
andreadevarda.bsky.social
Andrea de Varda
@andreadevarda.bsky.social
Postdoc at MIT BCS, interested in language(s) in humans and LMs

https://andrea-de-varda.github.io/
December 10, 2025 at 7:28 PM
I'd love to watch this, is there a recording?
November 21, 2025 at 4:11 PM
Why does this alignment emerge? There are similarities in how reasoning models and humans learn: first by observing worked examples (pretraining), then by practicing with feedback (RL). In the end, just like humans, they allocate more effort to harder problems. (6/6)
November 19, 2025 at 8:14 PM
Token count also captures differences across tasks. Avg. token count predicts avg. RT across domains (r = 0.97, left), and even item-level RTs across all tasks (r = 0.92 (!!), right). (5/6)
November 19, 2025 at 8:14 PM
We found that the number of reasoning tokens generated by the model reliably correlates with human RTs within each task (mean r = 0.57, all ps < .001). (4/6)
November 19, 2025 at 8:14 PM
Large reasoning models can solve many reasoning problems, but do their computations reflect how humans think?
We compared human RTs to DeepSeek-R1’s CoT length across seven tasks: arithmetic (numeric & verbal), logic (syllogisms & ALE), relational reasoning, intuitive reasoning, and ARC (3/6)
November 19, 2025 at 8:14 PM
Neural networks are powerful in-silico models for studying cognition: LLMs and CNNs already capture key behaviors in language and vision. But can they also capture the cognitive demands of human reasoning? (2/6)
November 19, 2025 at 8:14 PM