Josef Woldense
@woldense.bsky.social
4.3K followers 250 following 120 posts
josefwoldense.com
Posts Media Videos Starter Packs
woldense.bsky.social
Cool paper. I'm going to shamelessly plug my work here that also deals with LLMs for research

bsky.app/profile/wold...
woldense.bsky.social
Paper alert 📣

Rapid advances in AI has some believe that LLM agents can replace real participants in human-subject research. If true, this would be huge!

Following a growing body of research, we delve deeper into this topic and examine the merits of this claim.

🧵...

arxiv.org/abs/2509.03736
Are LLM Agents Behaviorally Coherent? Latent Profiles for Social Simulation
The impressive capabilities of Large Language Models (LLMs) have fueled the notion that synthetic agents can serve as substitutes for real participants in human-subject research. In an effort to evalu...
arxiv.org
Reposted by Josef Woldense
isanet.bsky.social
Preparing for your upcoming #ResearchTalk? Sign up for two courses, taught by @woldense.bsky.social, to enhance your #Communication and apply #Storytelling principles to your #Presenting, and #Teaching skills! Open to both ISA Members and non-members. Register: buff.ly/PUXowRN
Virtual PASS Course. The Research Presentation as Storytelling: A Two-Part Training. Part 1: October 9. Part 2: October 10. 11 AM to 3 PM (ET). Part 1 Cost: $35. Part 2 Cost: $35. ISA logo. Background: An open laptop with an open book sitting on the keyboard.
woldense.bsky.social
Congrats 🎉... looking forward to the coming research
Reposted by Josef Woldense
vincentab.bsky.social
Whoa—my book is up for pre-order!

𝐌𝐨𝐝𝐞𝐥 𝐭𝐨 𝐌𝐞𝐚𝐧𝐢𝐧𝐠: 𝐇𝐨𝐰 𝐭𝐨 𝐈𝐧𝐭𝐞𝐫𝐩𝐫𝐞𝐭 𝐒𝐭𝐚𝐭 & 𝐌𝐋 𝐌𝐨𝐝𝐞𝐥𝐬 𝐢𝐧 #Rstats 𝐚𝐧𝐝 #PyData

The book presents an ultra-simple and powerful workflow to make sense of ± any model you fit

The web version will stay free forever and my proceeds go to charity.

tinyurl.com/4fk56fc8
woldense.bsky.social
This looks fascinating!
babeheim.bsky.social
How to quantify the impact of AI on long-run cultural evolution? Published today, I give it a go!

400+ years of strategic dynamics in the game of Go (Baduk/Weiqi), from feudalism to AlphaGo!
Miyagawa Shuntei's 1898 painting, "Playing Go (Japanese Chess)"
woldense.bsky.social
There is more in the paper, but broadly speaking, our results identify a deceptive problem: surface-level plausibility masking deeper failure modes. Agents appear internally consistent while concealing systematic incoherence.

Be careful when using LLMs as human substitutes. They might fool you.
woldense.bsky.social
Take pairs where one of the agents has a preference of 1. Next, take pairs where one of the agents has a preference of 5. Now compare them. You can see pairs with a 1 have lower agreement scores than pairs with a 5. This is consistent across preference gaps
woldense.bsky.social
Let me give you another one.

If we both equally dislike soda, our common ground should lead to high agreement. Not so with our agents.
woldense.bsky.social
The problem persists, even when we try to guard against the problem of sycophancy (column 3 of the graph).

(see paper for more info on sycophancy)
woldense.bsky.social
Our estimate suggests that the suppression of disagreement is quite large. Our counterfactual agreements scores (expected in the graph) are significantly lower than the observed ones, and this is across preference gaps.

(see paper for info in mean shift)
woldense.bsky.social
To do this, we adopt a simplifying assumption – agents should disagree at the same rate as they agree. We already know one end of this spectrum -- the amount of agreement when agents are aligned (gap = 0). We establish the disagreement side (gap = 4), by assuming it to be the inverse of agreement
woldense.bsky.social
When agents are aligned, they reach close to the highest agreement score. Yet, when maximally different (gap = 4), they come nowhere near the lowest score. It seems agreement is amplified while disagreement is dampened.

Is it possible to estimate how much disagreement is being suppressed? Yes!
woldense.bsky.social
Looking at the graph, it appears consistent with our expectations, the more closely aligned the agents (smaller preference gap between agents), the higher the agreement score.

But there is a problem. Can you spot it?
woldense.bsky.social
What are the results? Are agents internally consistent?

At first glance, yes. After a more thorough analysis, the answer is no.
woldense.bsky.social
How do we measure agreement level?

With the aid of an LLM judge, we score each conversation (strongly disagree = 1 – strongly agree = 5). This yields a set of agreement scores for a given preference pair. Using bootstrap sampling, we derive the distribution of average agreement scores (range)
woldense.bsky.social
We elicit the agents’ preference on a topic (1-5 scale), then pair them in a conversation to see if they follow through on their preferences.

Expectation: The more closely agents align in their preferences, the more strongly they will agree. The further apart, the more they disagree.
woldense.bsky.social
The basic intuition of internal coherence: If a person says they strongly prefer water over soda, we expect them to follow through on it. When offered both, they should select water, not soda.

How do we test for internal coherence?
woldense.bsky.social
Unlike other studies that look at how successfully LLM agents adopt human personas, we take a different approach and ask: Once a persona has been adopted, are agents internally coherent?
woldense.bsky.social
Paper alert 📣

Rapid advances in AI has some believe that LLM agents can replace real participants in human-subject research. If true, this would be huge!

Following a growing body of research, we delve deeper into this topic and examine the merits of this claim.

🧵...

arxiv.org/abs/2509.03736
Are LLM Agents Behaviorally Coherent? Latent Profiles for Social Simulation
The impressive capabilities of Large Language Models (LLMs) have fueled the notion that synthetic agents can serve as substitutes for real participants in human-subject research. In an effort to evalu...
arxiv.org
woldense.bsky.social
Congrats!! I see more GETGOV workshops on the horizon
woldense.bsky.social
Unfortunately, no. I will make my way to Europe next year though, so hopefully we'll get a chance to catch up