Siyuan Song✈️COLM
@siyuansong.bsky.social
150 followers 320 following 34 posts
senior undergrad@UTexas Linguistics | @growai.bsky.social Looking for Ph.D position 26 Fall Comp Psycholing & CogSci, human-like AI, rock🎸 Prev: MIT BCS, VURI@Harvard Psych, Undergrad@SJTU Opinions are my own.
Posts Media Videos Starter Packs
Pinned
siyuansong.bsky.social
New preprint w/ @jennhu.bsky.social @kmahowald.bsky.social : Can LLMs introspect about their knowledge of language?
Across models and domains, we did not find evidence that LLMs have privileged access to their own predictions. 🧵(1/8)
Reposted by Siyuan Song✈️COLM
juand-r.bsky.social
Excited to present this at COLM tomorrow! (Tuesday, 11:00 AM poster session)
juand-r.bsky.social
One of the ways that LLMs can be inconsistent is the "generator-validator gap," where LLMs deem their own answers incorrect.

🎯 We demonstrate that ranking-based discriminator training can significantly reduce this gap, and improvements on one task often generalize to others!

🧵👇
A visualization of the generator-validator gap, where the LM likelihoods of for the generator and discriminator forms of questions are poorly correlated. Aligning the validator and generator rankings can fix it!
Reposted by Siyuan Song✈️COLM
sashaboguraev.bsky.social
I will be giving a short talk on this work at the COLM Interplay workshop on Friday (also to appear at EMNLP)!

Will be in Montreal all week and excited to chat about LM interpretability + its interaction with human cognition and ling theory.
sashaboguraev.bsky.social
A key hypothesis in the history of linguistics is that different constructions share underlying structure. We take advantage of recent advances in mechanistic interpretability to test this hypothesis in Language Models.

New work with @kmahowald.bsky.social and @cgpotts.bsky.social!

🧵👇!
Reposted by Siyuan Song✈️COLM
jessyjli.bsky.social
On my way to #COLM2025 🍁

Check out jessyli.com/colm2025

QUDsim: Discourse templates in LLM stories arxiv.org/abs/2504.09373

EvalAgent: retrieval-based eval targeting implicit criteria arxiv.org/abs/2504.15219

RoboInstruct: code generation for robotics with simulators arxiv.org/abs/2405.20179
siyuansong.bsky.social
Heading to #COLM2025 to present my first paper w/ @jennhu.bsky.social @kmahowald.bsky.social !

When: Tuesday, 11 AM – 1 PM
Where: Poster #75

Happy to chat about my work and topics in computational linguistics & cogsci!

Also, I'm on the PhD application journey this cycle!

Paper info 👇:
siyuansong.bsky.social
New preprint w/ @jennhu.bsky.social @kmahowald.bsky.social : Can LLMs introspect about their knowledge of language?
Across models and domains, we did not find evidence that LLMs have privileged access to their own predictions. 🧵(1/8)
Reposted by Siyuan Song✈️COLM
rtommccoy.bsky.social
🤖 🧠 NEW BLOG POST 🧠 🤖

What skills do you need to be a successful researcher?

The list seems long: collaborating, writing, presenting, reviewing, etc

But I argue that many of these skills can be unified under a single overarching ability: theory of mind

rtmccoy.com/posts/theory...
Illustration of the blog post's main argument, summarized as: "Theory of Mind as a Central Skill for Researchers: Research involves many skills.If each skill is viewed separately, each one takes a long time to learn. These skills can instead be connected via theory of mind – the ability to reason about the mental states of others. This allows you to transfer your abilities across areas, making it easier to gain new skills."
Reposted by Siyuan Song✈️COLM
kanishka.bsky.social
The compling group at UT Austin (sites.utexas.edu/compling/) is looking for PhD students!

Come join me, @kmahowald.bsky.social, and @jessyjli.bsky.social as we tackle interesting research questions at the intersection of ling, cogsci, and ai!

Some topics I am particularly interested in:
Picture of the UT Tower with "UT Austin Computational Linguistics" written in bigger font, and "Humans processing computers processing human processing language" in smaller font
Reposted by Siyuan Song✈️COLM
jessyjli.bsky.social
Can AI aid scientists amidst their own workflows, when they do not know step-by-step workflows and may not know, in advance, the kinds of scientific utility a visualization would bring?

Check out @sebajoe.bsky.social’s feature on ✨AstroVisBench:
nsfsimonscosmicai.bsky.social
Exciting news! Introducing AstroVisBench: A Code Benchmark for Scientific Computing and Visualization in Astronomy!

A new benchmark developed by researchers at the NSF-Simons AI Institute for Cosmic Origins is testing how well LLMs implement scientific workflows in astronomy and visualize results.
Reposted by Siyuan Song✈️COLM
Reposted by Siyuan Song✈️COLM
lampinen.bsky.social
Why does AI sometimes fail to generalize, and what might help? In a new paper (arxiv.org/abs/2509.16189), we highlight the latent learning gap — which unifies findings from language modeling to agent navigation — and suggest that episodic memory complements parametric learning to bridge it. Thread:
Latent learning: episodic memory complements parametric learning by enabling flexible reuse of experiences
When do machine learning systems fail to generalize, and what mechanisms could improve their generalization? Here, we draw inspiration from cognitive science to argue that one weakness of machine lear...
arxiv.org
Reposted by Siyuan Song✈️COLM
stefanfrank.bsky.social
Announcing the first (and perhaps only) Multilingual Minds and Machines Meeting! Come join us in Nijmegen, June 22-23, 2026, if you are interested in computational models of human multilingualism: mmmm2026.github.io
Reposted by Siyuan Song✈️COLM
catherinearnett.bsky.social
Did you know?

❌77% of language models on @hf.co are not tagged for any language
📈For 95% of languages, most models are multilingual
🚨88% of models with tags are trained on English

In a new blog post, @tylerachang.bsky.social and I dig into these trends and why they matter! 👇
Reposted by Siyuan Song✈️COLM
brendenlake.bsky.social
Our new lab for Human & Machine Intelligence is officially open at Princeton University!

Consider applying for a PhD or Postdoc position, either through Computer Science or Psychology. You can register interest on our new website lake-lab.github.io (1/2)
Reposted by Siyuan Song✈️COLM
Reposted by Siyuan Song✈️COLM
kmahowald.bsky.social
Can AI introspect? Surprisingly tricky to define what that means! And also interesting to test. New work from @siyuansong.bsky.social, @harveylederman.bsky.social, @jennhu.bsky.social and me on introspection in LLMs. See paper and thread for a definition and some experiments!
siyuansong.bsky.social
How reliable is what an AI says about itself? The answer depends on whether models can introspect. But, if an LLM says its temperature parameter is high (and it is!)….does that mean it’s introspecting? Surprisingly tricky to pin down. Our paper: arxiv.org/abs/2508.14802 (1/n)
Reposted by Siyuan Song✈️COLM
jennhu.bsky.social
Can AI models introspect? What does introspection even mean for AI?

We revisit a recent proposal by Comșa & Shanahan, and provide new experiments + an alternate definition of introspection.

Check out this new work w/ @siyuansong.bsky.social, @harveylederman.bsky.social, & @kmahowald.bsky.social 👇
siyuansong.bsky.social
How reliable is what an AI says about itself? The answer depends on whether models can introspect. But, if an LLM says its temperature parameter is high (and it is!)….does that mean it’s introspecting? Surprisingly tricky to pin down. Our paper: arxiv.org/abs/2508.14802 (1/n)
Reposted by Siyuan Song✈️COLM
harveylederman.bsky.social
exciting new paper from Siyuan! I really enjoyed working with him on this, inspired by important work by Murray Shanahan and Julia Comsa. Hard questions about how to operationalize the notion of “introspection” that’s relevant for practical applications in AI today. Hope you’ll check it out!
siyuansong.bsky.social
How reliable is what an AI says about itself? The answer depends on whether models can introspect. But, if an LLM says its temperature parameter is high (and it is!)….does that mean it’s introspecting? Surprisingly tricky to pin down. Our paper: arxiv.org/abs/2508.14802 (1/n)
siyuansong.bsky.social
Also check out our previous work that LMs do not introspect in grammaticality judgment and word prediction, to appear at COLM 2025: arxiv.org/pdf/2503.07513

And important work by Binder et al. that show evidence of privileged self-access in fine-tuned LLMs: openreview.net/forum?id=eb5...

(11/n)
arxiv.org
siyuansong.bsky.social
Taken together, our results suggest that: although LLMs can reason about the possible states of systems like themself, this does not imply that they have privileged self-access to their internal state - and this matters for introspection in AI systems.(10/n)
siyuansong.bsky.social
study2:we examined whether LLMs report their own temperature better than other models do. We found that self-reflection offers no advantage over temperature prediction (predicting based on the prompt and the generated text), whether within the same model or across different models.(9/n)
siyuansong.bsky.social
study1:we reproduced C&S’s temperature self-reporting case using a broader set of prompt and temperature settings. We found such self-reflection is highly sensitive to the prompt: even when the sampling temperature is low, a prompt 'generate a crazy sentence' leads to a high-temperature report.(8/n)
siyuansong.bsky.social
We performed two studies showing LLM failures to introspect per our definition. We think they illustrate some of the interesting subtleties in defining what AI introspection is in the relevant sense. (7/n)