Kanishk Gandhi
@gandhikanishk.bsky.social
830 followers 330 following 18 posts
PhD Student Stanford w/ Noah Goodman, studying reasoning, discovery, and interaction. Trying to build machines that understand people. StanfordNLP, Stanford AI Lab
Posts Media Videos Starter Packs
Pinned
gandhikanishk.bsky.social
1/13 New Paper!! We try to understand why some LMs self-improve their reasoning while others hit a wall. The key? Cognitive behaviors! Read our paper on how the right cognitive behaviors can make all the difference in a model's ability to improve with RL! 🧵
Reposted by Kanishk Gandhi
benpry.bsky.social
How can we combine the process-level insight that think-aloud studies give us with the large scale that modern online experiments permit? In our new CogSci paper, we show that speech-to-text models and LLMs enable us to scale up the think-aloud method to large experiments!
danielwurgaft.bsky.social
Excited to share a new CogSci paper co-led with @benpry.bsky.social!

Once a cornerstone for studying human reasoning, the think-aloud method declined in popularity as manual coding limited its scale. We introduce a method to automate analysis of verbal reports and scale think-aloud studies. (1/8)🧵
gandhikanishk.bsky.social
Can we record and study human chains of thought? Check out our new work led by @danielwurgaft.bsky.social and @benpry.bsky.social !!
danielwurgaft.bsky.social
Excited to share a new CogSci paper co-led with @benpry.bsky.social!

Once a cornerstone for studying human reasoning, the think-aloud method declined in popularity as manual coding limited its scale. We introduce a method to automate analysis of verbal reports and scale think-aloud studies. (1/8)🧵
Reposted by Kanishk Gandhi
gatodohq.bsky.social
Some absolutely marvellous work from @gandhikanishk.bsky.social et al! Wow!
gandhikanishk.bsky.social
12/13 Would also like to thank Charlie Snell, Dimitris Papailiopoulos, Eric Zelikman, Alex Havrilla, Rafael Rafaelov, @upiter.bsky.social and Archit Sharma for discussions about the magic and woes of RL training with LLMs.
gandhikanishk.bsky.social
11/13 Work with amazing collaborators Ayush Chakravarthy, Anikait Singh, Nathan Lile and @noahdgoodman.bsky.social
gandhikanishk.bsky.social
10/13 This paper gives us some clues as to what facilitated self-improvement in the recent generation of LLMs and what kind of data enables it. The key lies in exploration of the right behaviors!
gandhikanishk.bsky.social
9/13 Our findings reveal a fundamental connection between a model's initial reasoning behaviors and its capacity for improvement through RL. Models that explore verification, backtracking, subgoals, and backward chaining are primed for success.
gandhikanishk.bsky.social
8/13 By curating an extended pretraining set to amplify them, we enable Llama to match Qwen's improvement.
gandhikanishk.bsky.social
7/13 Can we apply these insights to pretraining? We analyze math pretraining sets like OpenWebMath & FineMath, finding these key behaviors are quite rare.
gandhikanishk.bsky.social
6/13 Empty and length matched empty chain-of-thought priming fails to produce improvement, reverting models to baseline performance. This shows it's the specific cognitive behaviors, not just longer outputs, enabling learning.
gandhikanishk.bsky.social
5/13 Crucially, the reasoning patterns matter more than having correct answers. Models primed with incorrect solutions that demonstrate the right cognitive behaviors still show substantial improvement. The behaviors are key.
gandhikanishk.bsky.social
4/13 We curate priming datasets with different behavior combinations and find that models primed with backtracking and verification consistently improve. Interestingly, RL selectively amplifies the most useful behaviors for reaching the goal.
gandhikanishk.bsky.social
3/13 Can we change a model's initial properties to enable improvement? Yes! After "priming" Llama, by finetuning on examples demonstrating these behaviors, it starts improving from RL just like Qwen. The priming jumpstarts the learning process.
gandhikanishk.bsky.social
2/13 We identify 4 key cognitive behaviors that enable successful learning: Verification (checking work), Backtracking (trying new approaches), Subgoal Setting (breaking problems down) & Backward Chaining (working backwards from a goal). Qwen naturally exhibits these, while Llama mostly lacks them.
gandhikanishk.bsky.social
1/13 New Paper!! We try to understand why some LMs self-improve their reasoning while others hit a wall. The key? Cognitive behaviors! Read our paper on how the right cognitive behaviors can make all the difference in a model's ability to improve with RL! 🧵
Reposted by Kanishk Gandhi
hadas.bsky.social
emotionally, i’m constantly walking into a glass door
Reposted by Kanishk Gandhi
matspike.bsky.social
Can Large Language Models THINK and UNDERSTAND? The answer from cognitive science is, of course, lolwut YES!

The more interesting question is CAN TOASTERS LOVE? Intriguingly, the answer is ALSO YES! And they love YOU
a romantic toaster presenting a single red rose
Reposted by Kanishk Gandhi
sungkim.bsky.social
They present a scientifically optimized recipe of “Pasta alla Cacio e pepe” based on their findings, enabling a consistently flawless execution of this classic dish.

"Phase behavior of Cacio and Pepe sauce"

arxiv.org/abs/2501.00536
gandhikanishk.bsky.social
These are actually good? No blatant physics violations at least? Definitely better than I expected
gandhikanishk.bsky.social
Actually can you try it with objects that it might have actually seen? Like a blue book falling on a tennis ball? I feel like in abstract prompts like these material properties are underspecified.
Reposted by Kanishk Gandhi
lampinen.bsky.social
What counts as in-context learning (ICL)? Typically, you might think of it as learning a task from a few examples. However, we’ve just written a perspective (arxiv.org/abs/2412.03782) suggesting interpreting a much broader spectrum of behaviors as ICL! Quick summary thread: 1/7
The broader spectrum of in-context learning
The ability of language models to learn a task from a few examples in context has generated substantial interest. Here, we provide a perspective that situates this type of supervised few-shot learning...
arxiv.org
gandhikanishk.bsky.social
I'll be at Neurips this week :) looking forward to catching up with folks! Please reach out if you want to chat!!
Reposted by Kanishk Gandhi
xuanalogue.bsky.social
Okay the people requested one so here is an attempt at a Computational Cognitive Science starter pack -- with apologies to everyone I've missed! LMK if there's anyone I should add!

go.bsky.app/KDTg6pv