Thomas Davidson
@thomasdavidson.bsky.social
840 followers 610 following 54 posts
Sociologist at Rutgers. Studies far-right politics, populism, and hate speech. Computational social science. https://www.thomasrdavidson.com/
Posts Media Videos Starter Packs
Reposted by Thomas Davidson
woahworkshop.bsky.social
Should WOAH start a mentorship programme? 🤔

As the workshop grows, reviewer expectations are rising.
We don’t want contributors from adjacent communities penalised by *CL norms.

Senior PhDs and beyond could be mentors.

Share your thoughts:

👉 forms.gle/safif3rU2rs5...
Reposted by Thomas Davidson
lauraknelson.bsky.social
📊The Sociology Department at Stony Brook University invites applications for *two* tenure-track Assistant Professor positions to begin in Fall 2026. [They] seek scholars who specialize in issues of global inequality and justice, broadly construed.

apply.interfolio.com/172669
Apply - Interfolio {{$ctrl.$state.data.pageTitle}} - Apply - Interfolio
apply.interfolio.com
Reposted by Thomas Davidson
thomasdavidson.bsky.social
Preparing my CSS class for next week and learned that Spotify put limits on its API that prevent most interesting queries developer.spotify.com/blog/2024-11...

This was my main go-to after the Twitter API shut down and now I have to find a new application. Does anyone have any other API reccs?
thomasdavidson.bsky.social
Preparing my CSS class for next week and learned that Spotify put limits on its API that prevent most interesting queries developer.spotify.com/blog/2024-11...

This was my main go-to after the Twitter API shut down and now I have to find a new application. Does anyone have any other API reccs?
thomasdavidson.bsky.social
Interested to read this more closely. One issue I worry about using LLMs for text analysis is the slippage between annotation and classification, particularly for zero-shot learning (is it annotation or classification?). Certainly appears to underscore the critical need for validation.
joachimbaumann.bsky.social
🚨 New paper alert 🚨 Using LLMs as data annotators, you can produce any scientific result you want. We call this **LLM Hacking**.

Paper: arxiv.org/pdf/2509.08825
We present our new preprint titled "Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation".
We quantify LLM hacking risk through systematic replication of 37 diverse computational social science annotation tasks.
For these tasks, we use a combined set of 2,361 realistic hypotheses that researchers might test using these annotations.
Then, we collect 13 million LLM annotations across plausible LLM configurations.
These annotations feed into 1.4 million regressions testing the hypotheses. 
For a hypothesis with no true effect (ground truth $p > 0.05$), different LLM configurations yield conflicting conclusions.
Checkmarks indicate correct statistical conclusions matching ground truth; crosses indicate LLM hacking -- incorrect conclusions due to annotation errors.
Across all experiments, LLM hacking occurs in 31-50\% of cases even with highly capable models.
Since minor configuration changes can flip scientific conclusions, from correct to incorrect, LLM hacking can be exploited to present anything as statistically significant.
Reposted by Thomas Davidson
lsemethodology.bsky.social
We're hiring an Assistant Professor in Computational Social Science ❗

📚 jobs.lse.ac.uk/Vacancies/W/...

Apply before 26 October and join an internationally outstanding group of social science methodologists 🌎
we're hiring assistant professor in computational social science, applications close 26/10/2025
Reposted by Thomas Davidson
fmerhout.bsky.social
Come join me in wonderful Copenhagen! 🇩🇰

My department is looking to fill at least two positions - any specialization and any level! The University of Copenhagen aims to be the best place for the best ideas. What’s yours?

Apply by Nov 15.
Call for two or more open-rank academic positions in Sociology
jobportal.ku.dk
thomasdavidson.bsky.social
Ha thanks, Rohan. It’s a short one but it scratches the surface of what we can do with LRMs. And I haven’t seen any evaluations like this in the CS literature. Let me know if you have any feedback
thomasdavidson.bsky.social
Substantively, the results show how reasoning effort and traces could help tasks like content moderation

There are, of course, caveats: LRMs do not replicate human cognition, the models have limited capabilities, and reasoning traces are not always faithful

Preprint: arxiv.org/pdf/2508.20262
arxiv.org
thomasdavidson.bsky.social
Analysis of the reasoning traces for Gemini 2.5 shows that the model identifies second-order factors when faced with these decisions, helping to address common false positives like flagging reclaimed slurs as hate speech (Warning: offensive language in example)
thomasdavidson.bsky.social
On a content moderation task, humans take longer and LRMs use more tokens when offensiveness is identical or fixed.

This suggests that LRM behavior is consistent with dual process theories of cognition, as the models expend more reasoning effort when simple heuristics are insufficient
thomasdavidson.bsky.social
The results are consistent across three frontier LRMs: o3, Gemini 2.5 Pro, and Grok 4
thomasdavidson.bsky.social
New pre-print on large reasoning models 🤖🧠

To what extent does LRM behavior resemble human reasoning processes?

I find that LRM reasoning effort predicts human decision time on a pairwise comparison task, and both humans and LRMs require more time/effort on challenging tasks
Reposted by Thomas Davidson
pengzell.bsky.social
WE ARE HIRING! 2 Lecturers in Quantitative Social Science. Want a friendly interdisciplinary department in one of the world's most vibrant cities? This just might be for you.

Apply by: 10 Oct

www.ucl.ac.uk/work-at-ucl/...
thomasdavidson.bsky.social
This is pretty much what the earlier work on debiasing word embeddings was doing. And it turns out that it doesn't work...

Cue one of the best CS paper titles aclanthology.org/N19-1061/
thomasdavidson.bsky.social
Thanks for reading this thread. The entire special issue can be found here: journals.sagepub.com/toc/smra/54/3

If you want to talk more about AI and sociology, I'll be at ASA and will be giving a talk on some ongoing AI research in the Saturday morning session on Culture and CSS
thomasdavidson.bsky.social
Of course, the methodological focus means that we don’t cover other important topics like impacts of AI on society and the politics of AI. There is a lot more work to be done.

See the recent Socius special issue for more perspectives on the sociology of AI journals.sagepub.com/topic/collec...
journals.sagepub.com
thomasdavidson.bsky.social
If you want to learn more about these articles, check out our editors’ introduction, where we provide an overview and discuss some central themes.

One point we emphasize is how the model ecosystem has matured & open-weight models are viable for many problems journals.sagepub.com/doi/10.1177/...
thomasdavidson.bsky.social
Stuhler, Ton, and Ollion show how LLMs enable more complex information extraction tasks that can be applied to text corpora, with an application to the study of obituaries journals.sagepub.com/doi/abs/10.1...
thomasdavidson.bsky.social
A study by Than et al. revisits an earlier SMR paper, exploring how LLMs can be used for qualitative coding tasks, providing detailed guidance on how to approach the problem journals.sagepub.com/doi/full/10....
thomasdavidson.bsky.social
Law and Roberto showcase how vision language models like GPT-4o can extract information from satellite images, applying these techniques to study segregation in the built environment journals.sagepub.com/doi/full/10....