Tanise Ceron
@taniseceron.bsky.social
120 followers 140 following 23 posts
Postdoc @milanlp.bsky.social
Posts Media Videos Starter Packs
Pinned
taniseceron.bsky.social
📣 New Preprint!
Have you ever wondered what the political content in LLM's training data is? What are the political opinions expressed? What is the proportion of left- vs right-leaning documents in the pre- and post-training data? Do they correlate with the political biases reflected in models?
taniseceron.bsky.social
Great collaboration with Dmitry Nikolaev, @dominsta.bsky.social and @deboranozza.bsky.social ☺️
taniseceron.bsky.social
- Finally, and for me, most interestingly, our analysis suggests that political biases are already encoded during the pre-training stage.

Taken these evidences together, we highlight important implications these results play on data processing in the development of fairer LLMs.
taniseceron.bsky.social
- There's a strong correlation (Pearson r=0.90) between the predominant stances in the training data and the models’ behavior when probed for political bias on eight policy issues (e.g., environmental protection, migration, etc).
taniseceron.bsky.social
- Source domains of pre-training documents differ significantly, with right-leaning content containing twice as many blog posts and left-leaning content 3 times as many news outlets.
taniseceron.bsky.social
- The framing of political topics varies considerably: right-leaning labeled documents prioritize stability, sovereignty, and cautious reform via technology or deregulation, while left-leaning documents emphasize urgent, science-led mobilization for systemic transformation and equity.
taniseceron.bsky.social
- left-leaning documents consistently outnumber right-leaning ones by a factor of 3 to 12 across training datasets.
- pre-training corpora contains about 4 times more politically engaged content than post-training data.
taniseceron.bsky.social
We have the answers of these questions here : arxiv.org/pdf/2509.22367

We analyze the political content of the training data from OLMO2, the largest fully open-source model.
🕵️‍♀️ We run an analysis in all the datasets (2 pre- and 2 post-training) used to train the models. Here are our findings:
arxiv.org
taniseceron.bsky.social
📣 New Preprint!
Have you ever wondered what the political content in LLM's training data is? What are the political opinions expressed? What is the proportion of left- vs right-leaning documents in the pre- and post-training data? Do they correlate with the political biases reflected in models?
Reposted by Tanise Ceron
cscl-bot.bsky.social
Tanise Ceron, Dmitry Nikolaev, Dominik Stammbach, Debora Nozza: What Is The Political Content in LLMs' Pre- and Post-Training Data? https://arxiv.org/abs/2509.22367 https://arxiv.org/pdf/2509.22367 https://arxiv.org/html/2509.22367
taniseceron.bsky.social
Thanks SoftwareCampus for supporting Multiview, the organizers of INRA, and Sourabh Dattawad and @agnesedaff.bsky.social for the great collaboration!
taniseceron.bsky.social
Our evaluation with normative metrics shows that this approach does not diversify only frames in user's history, but also sentiment and news categories. These findings demonstrate that framing acts as a control lever for enhancing normative diversity.
taniseceron.bsky.social
In this paper, we propose introduce media frames as a device for diversifying perspectives in news recommenders. Our results show an improvement in exposure to previously unclicked frames up to 50%.
taniseceron.bsky.social
Today Sourabh Dattawad presented our work "Leveraging Media Frames to Improve Normative Diversity in News Recommendations" at INRA (International Workshop on News Recommendation and Analytics) co-located with RecSys 2025 in Prague.
arxiv.org/pdf/2509.02266
Reposted by Tanise Ceron
joachimbaumann.bsky.social
🚨 New paper alert 🚨 Using LLMs as data annotators, you can produce any scientific result you want. We call this **LLM Hacking**.

Paper: arxiv.org/pdf/2509.08825
We present our new preprint titled "Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation".
We quantify LLM hacking risk through systematic replication of 37 diverse computational social science annotation tasks.
For these tasks, we use a combined set of 2,361 realistic hypotheses that researchers might test using these annotations.
Then, we collect 13 million LLM annotations across plausible LLM configurations.
These annotations feed into 1.4 million regressions testing the hypotheses. 
For a hypothesis with no true effect (ground truth $p > 0.05$), different LLM configurations yield conflicting conclusions.
Checkmarks indicate correct statistical conclusions matching ground truth; crosses indicate LLM hacking -- incorrect conclusions due to annotation errors.
Across all experiments, LLM hacking occurs in 31-50\% of cases even with highly capable models.
Since minor configuration changes can flip scientific conclusions, from correct to incorrect, LLM hacking can be exploited to present anything as statistically significant.
Reposted by Tanise Ceron
milanlp.bsky.social
Last week we held our 1st MilaNLP retreat by beautiful Lago Maggiore! ⛰️🌊
We shared research ideas, stories (academic & beyond), and amazing food. It was a great time to connect outside of the usual lab working days, and most importantly, strengthen our bonds as a team. #ResearchLife #NLProc
Reposted by Tanise Ceron
bsavoldi.bsky.social
🔍 Stiamo studiando come l'AI viene usata in Italia e per farlo abbiamo costruito un sondaggio!

👉 bit.ly/sondaggio_ai...

(è anonimo, richiede ~10 minuti, e se partecipi o lo fai girare ci aiuti un sacco🙏)

Ci interessa anche raggiungere persone che non si occupano e non sono esperte di AI!
Qualtrics Survey | Qualtrics Experience Management
The most powerful, simple and trusted way to gather experience data. Start your journey to experience management and try a free account today.
bit.ly
taniseceron.bsky.social
Reminder for the importance of evaluating political biases robustly. :)
milanlp.bsky.social
#MemoryMonday #NLProc Beyond Prompt Brittleness: Evaluating the Reliability and Consistency of Political Worldviews in LLMs" (@taniseceron.bsky.social et al.,) evaluate the consistency of political worldviews in LLMs, unveiling fine-grained stances in policy issues.
Beyond Prompt Brittleness: Evaluating the Reliability and Consistency of Political Worldviews in LLMs
Tanise Ceron, Neele Falk, Ana Barić, Dmitry Nikolaev, Sebastian Padó. Transactions of the Association for Computational Linguistics, Volume 12. 2024.
aclanthology.org
Reposted by Tanise Ceron
dirkhovy.bsky.social
We (w/ @diyiyang.bsky.social, @zhuhao.me, & Bodhisattwa Prasad Majumder) are excited to present our #NAACL25 tutorial on Social Intelligence in the Age of LLMs!
It will highlight long-standing and emerging challenges of AI interacting w humans, society & the world.
⏰ May 3, 2:00pm-5:30pm Room Pecos
Reposted by Tanise Ceron
verenakunz.bsky.social
Join us in an hour at 17:00 (CEST) for @taniseceron.bsky.social's talk on "Evaluating Political Bias: Insights into Robustness and Multilinguality“. Access to Zoom at join.slack.com/t/tadapolisc... or send me a ✉️
verenakunz.bsky.social
The #TaDa Speaker Series is back for the spring 🎉 We're looking forward to an exciting line-up of talks by @prashantgarg.bsky.social, @miriamschirmer.bsky.social, @chdausgaard.bsky.social, @taniseceron.bsky.social, @lukashetzer.bsky.social, and Catarina Pereira! More infos at tada.cool & on Slack ⬇️
The flyer shows the overview of the programme for the Text-as-Data Spring Term Speaker Series 2025, which usually takes place on Wednesdays at 5pm Berlin time.
taniseceron.bsky.social
Sure, it's here: github.com/tceron/eval_...
The code mapping is in the readme file. :)
github.com
Reposted by Tanise Ceron
cklamm.bsky.social
🥁 It's the second half of our 🌱 speaker series (tada.cool) this term, and we couldn't be more excited! Next week (Wednesday, April 30 at 5pm CET), we have the pleasure of welcoming @taniseceron.bsky.social to share insights on "Facilitating Information Access Through Language Models". More details ⬇️
verenakunz.bsky.social
The #TaDa Speaker Series is back for the spring 🎉 We're looking forward to an exciting line-up of talks by @prashantgarg.bsky.social, @miriamschirmer.bsky.social, @chdausgaard.bsky.social, @taniseceron.bsky.social, @lukashetzer.bsky.social, and Catarina Pereira! More infos at tada.cool & on Slack ⬇️
The flyer shows the overview of the programme for the Text-as-Data Spring Term Speaker Series 2025, which usually takes place on Wednesdays at 5pm Berlin time.
taniseceron.bsky.social
change the political worldviews of models. In our study, we find that the previous version (Llama-2) consistently reflects more left-leaning views. However, it does depend on the policy issue as we found clear stances of the models only towards social state welfare, environment protection and [2/3]
taniseceron.bsky.social
I agree 100% that we need to understand what they're measuring, and specifically, how they're aligning the models to be hold certain types of political worldviews. However, I find your results rather puzzling because Llama3.1 was released much before they started announcing their strategy to [1/3]