Daniel Chechelnitsky
@dchechel.bsky.social
37 followers 86 following 9 posts
PhDing @ CMU LTI
Posts Media Videos Starter Packs
Pinned
dchechel.bsky.social
What if AI played the role of your sassy gay bestie 🏳️‍🌈 or AAVE-speaking friend 👋🏾?

You: “Can you plan a trip?”
🤖 AI: “Yasss queen! let’s werk this babe✨💅”

LLMs can talk like us, but it shapes how we trust, rely on & relate to them 🧵

📣 our #FAccT2025 paper: bit.ly/3HJ6rWI

[1/9]
Reposted by Daniel Chechelnitsky
taolongg.bsky.social
tomorrow 6/20, i'm presenting this paper at #alt_FAccT, a NYC local meeting for @FAccTConference

✨🎤 paper session #3 🎤✨
🗽1:30pm June 20, Fri @ MSR NYC🗽

⬇️ our #FAccT2025 paper is abt “what if ur ChatGPT spoke queer slang and AAVE?”

📚🔗 bit.ly/not-like-us
dchechel.bsky.social
And here are some examples where users enjoyed the interaction with the sociolectal LLMs:

😊 “It just sounds more fun to interact with” -AAE participant

💅 “I enjoy being called a diva!” -Queer slang participant

[8/9]
dchechel.bsky.social
Lastly we asked users for justifications for their LLM preference. Here are a few comments about the sociolect LLMs:

🚫“Agent [AAELM] using AAE sounds like a joke and not natural.” -AAE participant

🚫“Even people who use LGBTQ slang don’t talk like that constantly...” -Queer slang participant

[7/9]
dchechel.bsky.social
We also were curious into seeing how each of the user perceptions impacted user reliance on LLMs. For this we observed that generally, perception variables were positively associated with reliance. 😄

[6/9]
dchechel.bsky.social
We also observed user perceptions: trust, social proximity, satisfaction, frustration, and explicit preference for an LLM using sociolects.

Notably, we notice how AAE participants explicitly preferred the SAELM over the AAELM, whereas this wasn’t the case for Queer slang participants. 💙💚

[5/9]
dchechel.bsky.social
In our study we find that AAE users rely more on the SAE LLM over AAELM, while for Queer slang users there is no difference between the SAE LLM and QSLM.

This shows that for some sociolects, users will rely more on an LLM in Standard English than one in a sociolect they use themselves. 🤎🩷

[4/9]
dchechel.bsky.social
We run two parallel studies:

1: with AAE speakers using AAE LLM (AAELM) 👋🏾
2: with Queer slang speakers using Queer slang LLM (QSLM) 🏳️‍🌈

In each, participants watched videos and were offered to use either a Standard English LLM or AAELM/QSLM to help answer questions.

[3/9]
dchechel.bsky.social
Our study (n=985) looks at how AAVE speakers and Queer slang speakers perceive and rely on LLMs’ use of their sociolect (i.e., a dialect centered around a social class). 🗣️

This answers our main research question:

“How do users behave and feel when engaging with a sociolectal LLM?” 🤷🏻🤷🏾‍♀️🤷🏽‍♂️

[2/9]
dchechel.bsky.social
What if AI played the role of your sassy gay bestie 🏳️‍🌈 or AAVE-speaking friend 👋🏾?

You: “Can you plan a trip?”
🤖 AI: “Yasss queen! let’s werk this babe✨💅”

LLMs can talk like us, but it shapes how we trust, rely on & relate to them 🧵

📣 our #FAccT2025 paper: bit.ly/3HJ6rWI

[1/9]
Reposted by Daniel Chechelnitsky
shaily99.bsky.social
🖋️ Curious how writing differs across (research) cultures?
🚩 Tired of “cultural” evals that don't consult people?

We engaged with interdisciplinary researchers to identify & measure ✨cultural norms✨in scientific writing, and show that❗LLMs flatten them❗

📜 arxiv.org/abs/2506.00784

[1/11]
An overview of the work “Research Borderlands: Analysing Writing Across Research Cultures” by Shaily Bhatt, Tal August, and Maria Antoniak. The overview describes that We  survey and interview interdisciplinary researchers (§3) to develop a framework of writing norms that vary across research cultures (§4) and operationalise them using computational metrics (§5). We then use this evaluation suite for two large-scale quantitative analyses: (a) surfacing variations in writing across 11 communities (§6); (b) evaluating the cultural competence of LLMs when adapting writing from one community to another (§7).
Reposted by Daniel Chechelnitsky
jordant.bsky.social
🏳️‍🌈🎨💻📢 Happy to share our workshop study on queer artists’ experiences critically engaging with GenAI

Looking forward to presenting this work at #FAccT2025 and you can read a pre-print here:
arxiv.org/abs/2503.09805
Academic paper titled un-straightening generative ai: how queer artists surface and challenge the normativity of generative ai models

The piece is written by Jordan Taylor, Joel Mire, Franchesca Spektor, Alicia DeVrio, Maarten Sap, Haiyi Zhu, and Sarah Fox.

As an image titled 24 attempts at intimacy showing 24 ai generated images with the word intimacy, none of which seems to include same gender couples
Reposted by Daniel Chechelnitsky
maartensap.bsky.social
RLHF is built upon some quite oversimplistic assumptions, i.e., that preferences between pairs of text are purely about quality. But this is an inherently subjective task (not unlike toxicity annotation) -- so we wanted to know, do biases similar to toxicity annotation emerge in reward models?
joelmire.bsky.social
Reward models for LMs are meant to align outputs with human preferences—but do they accidentally encode dialect biases? 🤔

Excited to share our paper on biases against African American Language in reward models, accepted to #NAACL2025 Findings! 🎉

Paper: arxiv.org/abs/2502.12858 (1/10)
Screenshot of Arxiv paper title, "Rejected Dialects: Biases Against African American Language in Reward Models," and author list: Joel Mire, Zubin Trivadi Aysola, Daniel Chechelnitsky, Nicholas Deas, Chrysoula Zerva, and Maarten Sap.
Reposted by Daniel Chechelnitsky
joelmire.bsky.social
Reward models for LMs are meant to align outputs with human preferences—but do they accidentally encode dialect biases? 🤔

Excited to share our paper on biases against African American Language in reward models, accepted to #NAACL2025 Findings! 🎉

Paper: arxiv.org/abs/2502.12858 (1/10)
Screenshot of Arxiv paper title, "Rejected Dialects: Biases Against African American Language in Reward Models," and author list: Joel Mire, Zubin Trivadi Aysola, Daniel Chechelnitsky, Nicholas Deas, Chrysoula Zerva, and Maarten Sap.
Reposted by Daniel Chechelnitsky
ltiatcmu.bsky.social
Looking for all your LTI friends on Bluesky? The LTI Starter Pack is here to help!

go.bsky.app/NhTwCVb