Paper: arxiv.org/pdf/2509.08825
@seanjwestwood.bsky.social.
@seanjwestwood.bsky.social.
We have a new paper - led by Desheng Hu, now accepted at @icwsm.bsky.social - exploring that and finding many issues
Preprint: arxiv.org/abs/2511.12920
🧵👇
We have a new paper - led by Desheng Hu, now accepted at @icwsm.bsky.social - exploring that and finding many issues
Preprint: arxiv.org/abs/2511.12920
🧵👇
@joachimbaumann.bsky.social, who will present co-authored work on "Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation". Paper and information on how to join ⬇️
@joachimbaumann.bsky.social, who will present co-authored work on "Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation". Paper and information on how to join ⬇️
The promise is revolutionary for science & policy. But there’s a huge "IF": Do these simulations actually reflect reality?
To find out, we introduce SimBench: The first large-scale benchmark for group-level social simulation. (1/9)
The promise is revolutionary for science & policy. But there’s a huge "IF": Do these simulations actually reflect reality?
To find out, we introduce SimBench: The first large-scale benchmark for group-level social simulation. (1/9)
✓ LLMs are brittle data annotators
✓ Downstream conclusions flip frequently: LLM hacking risk is real!
✓ Bias correction methods can help but have trade-offs
✓ Use human expert whenever possible
✓ LLMs are brittle data annotators
✓ Downstream conclusions flip frequently: LLM hacking risk is real!
✓ Bias correction methods can help but have trade-offs
✓ Use human expert whenever possible
Paper: arxiv.org/pdf/2509.08825
Last week, we -- the (Amazing) Social Computing Group, held an internal hackathon to work on our informally called “Cultural Imperialism” project.
Last week, we -- the (Amazing) Social Computing Group, held an internal hackathon to work on our informally called “Cultural Imperialism” project.
Experiment + *evidence-based* mitigation strategies in this preprint 👇
Paper: arxiv.org/pdf/2509.08825
Experiment + *evidence-based* mitigation strategies in this preprint 👇
Paper: arxiv.org/pdf/2509.08825
Paper: arxiv.org/pdf/2509.08825
Just joined @milanlp.bsky.social as a Postdoc, working with the amazing @dirkhovy.bsky.social on large language models and computational social science!
Just joined @milanlp.bsky.social as a Postdoc, working with the amazing @dirkhovy.bsky.social on large language models and computational social science!