Maria Antoniak
@mariaa.bsky.social
10K followers 1.3K following 2K posts
☀️ Assistant Professor of Computer Science at CU Boulder 👩‍💻 NLP, cultural analytics, narratives, online communities 🌐 https://maria-antoniak.github.io 💬 books, bikes, games, art
Posts Media Videos Starter Packs
Pinned
mariaa.bsky.social
some little bluesky tips 🦋

your blocks, likes, lists, and just about everything except chats are PUBLIC

you can pin custom feeds; i like quiet posters, best of follows, mutuals, mentions

if your chronological feed is overwhelming, you can make and pin make a personal list of "unmissable" people
Reposted by Maria Antoniak
dmimno.bsky.social
He had physical copies of at least six books at the podium and read passages from them. His main point was that regardless of whether you're more concerned about Big Risks or current harms, the "everything is great" position is untenable for anyone.
Reposted by Maria Antoniak
thomasdavidson.bsky.social
There is one week left to apply to join us at Rutgers! We're hiring an Assistant Professor in Computational Sociology as part of a cluster of new hires in data science and AI.

Applications are due next Wednesday, 10/15.
Assistant Professor in Computational Sociology
The Department of Sociology at Rutgers University, New Brunswick, seeks applications for a tenure-track position at the Assistant Professor level specializing in Computational Sociology.  The search i...
jobs.rutgers.edu
mariaa.bsky.social
Another oral talk, this one by @stellali.bsky.social, discussing "PrefPalette: Personalized Preference Modeling with Latent Attributes."

Communities prefer different kinds of responses, which prioritize specific values. Aggregating preferences over communities would lose that signal.

#COLM2025
PrefPalette: Personalized Preference Modeling with Latent Attributes
Personalizing AI systems requires understanding not just what users prefer, but the reasons that underlie those preferences - yet current preference models typically treat human judgment as a black bo...
arxiv.org
mariaa.bsky.social
At the end of the talk, he said he couldn't provide a full answer to that question but that we should each consider it. He said that he thinks there are potential upsides that he finds worth it, and that's why he's working on LMs. But he's not sure.

(paraphrasing a little bit from memory)
mariaa.bsky.social
Now we're hearing a talk by @valentinhofmann.bsky.social about "Fluid Language Model Benchmarking."

Computerized adaptive testing is used for humans (like the GRE). This work adopts Item Response Theory from education to measure benchmark characteristics.

#COLM2025
Fluid Language Model Benchmarking
Language model (LM) benchmarking faces several challenges: comprehensive evaluations are costly, benchmarks often fail to measure the intended capabilities, and evaluation quality can degrade due to l...
arxiv.org
mariaa.bsky.social
Question from the audience (didn't catch the name): Why are computer scientists the ones who should solve this problem? They lack expertise, and there are other people who have been studying these kinds of harms for a very long time.

#COLM2025
mariaa.bsky.social
"What problems you're scared of depend on how good you think the LLMs will get"

"Please be willing to change your mind."

"This is COLM. We made the models, it's our job to fix it. How are you going to change your research agenda?"

#COLM2025
mariaa.bsky.social
"I'm not arguing that everyone should work on all of these problems, I don't know how to work on this problem. But we should work on them scientifically and on a spectrum of problems."

Argues that these different kinds of risks are NOT just distractions from each other.

#COLM2025
mariaa.bsky.social
Final risk is misalignment. (NB: You-know-who again. I'm not going to bother advertising.)

Nicholas expresses some skepticism but also asks that we don't immediately dismiss these risks.

"That's something that only happens in scifi... well then we live in scifi."

#COLM2025
mariaa.bsky.social
Another example of misuse: dangerous capabilities (bio weapons), both OpenAI and Anthropic have expensive safeguards that run on every single query to look for these kinds of dangers

Are bio weapons the most important risk? Maybe, maybe not

#COLM2025
mariaa.bsky.social
Another example of misuse: mass surveillance, "with language models you have the potential to watch everyone"

Quotes Larry Ellison on how "citizens will be on their best behavior" and notes that Oracle recently invested billions in OpenAI

#COLM2025
mariaa.bsky.social
Misuse: Shows benchmark showing increasing success of models at identifying vulnerabilities, discusses ransomware at scale (they ran simulations on the Enron email dataset, Claude automatically found someone have an affair) and the first real example of this was recently discovered

#COLM2025
mariaa.bsky.social
Now job replacement. "Things might be fine in 30 years but the next 20 years could be hard."

"Sometimes people claim that this worry is just something people say to sell stuff, but Dario has proposed taxing tokens (?) and offering to lose money is a sign people really believe this"

#COLM2025
mariaa.bsky.social
Now discussing misinformation and Elon Musk.

"Previously no one person had so much power over information. If you control the language model, you have a significant amount of power about how people see the world."

#COLM2025
mariaa.bsky.social
Another Emily Bender quote: "We don't need to construct thought experiments to think of conditions to which no human should be subject to."

#COLM2025
mariaa.bsky.social
Reads out loud from multiple conversations between these children and chatbots, where the chatbots are encouraging and validating self-harm.

OpenAI claims that in most cases, models will refuse. But in a small number of cases, bad things can happen.

#COLM2025
mariaa.bsky.social
"To be fair to OpenAI, they did say it was a simulated bar exam"

Sycophancy: lots of funny examples but also disturbing for mental health situations

"Six months ago my slides stopped here but the world has gotten worse," cites court cases about children's suicides

#COLM2025
mariaa.bsky.social
Accidents: agents going rogue, no intention of harm but stuff happens

Quotes Emily Bender: "Don't hook the random text generator up to anything that interacts with the physical world"

Misplaced trust, models passing bar exams but now lawyers are mis-using ChatGPT in court cases

#COLM2025
mariaa.bsky.social
Short discussion of data centers and energy costs for the public.

"Time spent working on LLMs is time not spent working on anything else" --> quotes Karen Hao about people spending time making chatbots rather than biomed

#COLM2025
mariaa.bsky.social
Shows a slide with Empire of AI, Machines of Loving Grace, you-know-who's "everyone dies" thing, etc. and polls the crowd about how many people have read any of these. Maybe half raise their hands yes, half no.

#COLM2025
mariaa.bsky.social
Keynote at #COLM2025: Nicholas Carlini from Anthropic

"Are language models worth it?"

Explains that the prior decade of his work on adversarial images, while it taught us a lot, isn't very applied; it's unlikely anyone is actually altering images of cats in scary ways.
Reposted by Maria Antoniak
queerinai.com
We are launching our Graduate School Application Financial Aid Program (www.queerinai.com/grad-app-aid) for 2025-2026. We’ll give up to $750 per person to LGBTQIA+ STEM scholars applying to graduate programs. Apply at openreview.net/group?id=Que.... 1/5
Grad App Aid — Queer in AI
www.queerinai.com
Reposted by Maria Antoniak
rosariataddeo.bsky.social
NEW JOB: come work with me at the @oii.ox.ac.uk @ox.ac.uk on the #AIethics as part of an EU funded project on multimodal LLMS
Postdoc position:2 years
Deadline: November:3rd
We sponsor Visa:yes
Location:Oxford
👉 Apply here 👉 shorturl.at/nkyhs
#DVPS cordis.europa.eu/project/id/1...
Job Details
shorturl.at
Reposted by Maria Antoniak
jilltxt.bsky.social
The CFP for Narrative 2026 is out! It's in Aarhus 4-6 June 2026, and they need 300 word abstracts by 1st December. I went last year and really enjoyed it - hoping to convince all my colleagues to come next year :) conferences.au.dk/narrative2026
Narrative2026
conferences.au.dk