Clara Na
@clarana.bsky.social
2.2K followers 390 following 17 posts
PhD student @ CMU LTI. efficiency/data in NLP/ML
Posts Media Videos Starter Packs
Pinned
clarana.bsky.social
Building/customizing your own LLM? You'll want to curate training data for it, but how do you know what makes the data good?
You can try out recipes👩‍🍳 iterate on ✨vibes✨ but we can't actually test all possible combos of tweaks,,, right?? 🙅‍♂️WRONG! arxiv.org/abs/2410.15661 (1/n) 🧵
clarana.bsky.social
Yes! tbh this method is probably much more immediately useful for helping one understand subtle differences between [models trained on] subtly different data subsets, vs a loftier goal of helping one find "the" best data mixture -- to anyone considering this method, please feel free to reach out :)
tedunderwood.com
The method in this paper was designed to find an optimal data mixture. But researchers in the human sciences who are training models *in order to understand the effect of the data* might also consider this as a clever way of evaluating hundreds of subsets without training hundreds of models. #MLSky
Figure showing a modular training strategy for evaluating domain importance in training data.
At the top, a question is posed: “Which domain is most beneficial to add to the training data?” Below, the left panel labeled Modular Training displays colored blocks representing separate models trained on distinct data partitions. Each block corresponds to a “base unit” of data, and blocks of different colors represent different domains. The right panel labeled Evaluation shows overlapping combinations of these trained models being evaluated together. The strategy allows for reuse of modularly trained models and performs evaluation on parameter averages, enabling efficient simulation of many data mixtures without retraining full models for each. A legend at the bottom explains that each block represents one model trained on x billion tokens, and each outlined group represents one evaluation.
clarana.bsky.social
I almost never use these so I always thought that they were cute little things that let seatmates watch the same movie
clarana.bsky.social
Come through! #492 in Hall 2!, 10am-12:30pm
Reposted by Clara Na
Reposted by Clara Na
jacobcares.bsky.social
I'm in Singapore for @iclr-conf.bsky.social ! Come check out our spotlight paper on the environmental impact of training OLMo (link in next tweet) during the Saturday morning poster session from 10-12:30 -- happy to chat about this or anything else! DMs should be open, email works too
Reposted by Clara Na
datarescueproject.org
We've received multiple notes that NOAA research services (Office of Oceanic and Atmospheric Research) may go offline at midnight. @safeguardingdata.bsky.social is working on web archiving, but if others want to nominate on this, that might be good: digital2.library.unt.edu/nomination/G...
Nomination Tool: Project URL Nomination
digital2.library.unt.edu
Reposted by Clara Na
uhleeeeeeeshuh.bsky.social
How can we better think and talk about human-like qualities attributed to language technologies like LLMs? In our #CHI2025 paper, we taxonomize how text outputs from cases of user interactions with language technologies can contribute to anthropomorphism. arxiv.org/abs/2502.09870 1/n
Image of the first page of the CHI 2025 paper titled "A Taxonomy of Linguistic Expressions That Contribute To Anthropomorphism of Language Technologies" by authors Alicia DeVrio, Myra Cheng, Lisa Egede, Alexandra Olteanu, & Su Lin Blodgett
Reposted by Clara Na
akhilayerukola.bsky.social
Did you know? Gestures used to express universal concepts—like wishing for luck—vary DRAMATICALLY across cultures?
🤞means luck in US but deeply offensive in Vietnam 🚨

📣 We introduce MC-SIGNS, a test bed to evaluate how LLMs/VLMs/T2I handle such nonverbal behavior!

📜: arxiv.org/abs/2502.17710
Figure showing that interpretations of gestures vary dramatically across regions and cultures. ‘Crossing your fingers,’ commonly used in the US to wish for good luck, can be deeply offensive to female audiences in parts of Vietnam. Similarly, the 'fig gesture,' a playful 'got your nose' game with children in the US, carries strong sexual connotations in Japan and can be highly offensive.
Reposted by Clara Na
kylelo.bsky.social
the science of LMs should be fully open✨

today @akshitab.bsky.social @natolambert.bsky.social and I are giving our #neurips2024 tutorial on language model development.

everything from data, training, adaptation. published or not, no secrets 🫡

tues, 12/10, 9:30am PT ☕️

neurips.cc/virtual/2024...
NeurIPS Tutorial Opening the Language Model Pipeline: A Tutorial on Data Preparation, Model Training, and AdaptationNeurIPS 2024
neurips.cc
Reposted by Clara Na
casilli.bsky.social
How open is “open” AI, really?
It isn’t just about making models reusable. If the origin of data is opaque, if labor is hidden & exploited, if frameworks are dominated by Big Tech, if computational power is mastered by an oligopoly…‘open’ is just a label.

Meredith Whittaker & friends in Nature.
Reposted by Clara Na
marcmarone.com
I noticed a lot of starter packs skewed towards faculty/industry, so I made one of just NLP & ML students: go.bsky.app/vju2ux

Students do different research, go on the job market, and recruit other students. Ping me and I'll add you!
Reposted by Clara Na
lindiatjuatja.bsky.social
💬 Have you or a loved one compared LM probabilities to human linguistic acceptability judgments? You may be overcompensating for the effect of frequency and length!
🌟 In our new paper, we rethink how we should be controlling for these factors 🧵:
Screenshot of the paper title "What Goes Into a LM Acceptability Judgment? Rethinking the Impact of Frequency and Length"
clarana.bsky.social
Hi I am at 232 in the back of the riverfront room!
clarana.bsky.social
I'm at EMNLP! Presenting the poster for this paper on Thursday morning (10:30-12), Session F Riverfront Hall, come say hi :)
clarana.bsky.social
Building/customizing your own LLM? You'll want to curate training data for it, but how do you know what makes the data good?
You can try out recipes👩‍🍳 iterate on ✨vibes✨ but we can't actually test all possible combos of tweaks,,, right?? 🙅‍♂️WRONG! arxiv.org/abs/2410.15661 (1/n) 🧵
Reposted by Clara Na
lindiatjuatja.bsky.social
(Hehe first bsky post!) I'll be at #EMNLP2024 💃🌴! Happy to chat about (among other things):
✨linguistically+cognitively motivated evaluation
✨NLP for low-resource+endangered languages
✨figuring out what features of language data LMs are *actually* learning
I'll be presenting two posters 🧵:
clarana.bsky.social
scrolling,,, minimal doom ?!
Reposted by Clara Na
mariaa.bsky.social
A starter pack for #NLP #NLProc researchers! 🎉

go.bsky.app/SngwGeS
clarana.bsky.social
I'll be presenting our paper at #EMNLP2024 next week -- see y'all in Miami🌴! This was my Summer 2023 work @ai2.bsky.social Grateful to my wonderful collaborators @ianmagnusson.bsky.social @ananyahjha93.bsky.social @tomsherborne.bsky.social & mentors @strubell.bsky.social Jesse, and Pradeep (6/n)
clarana.bsky.social
We can even predict larger model perplexity scores w/ smaller model proxy evals, AND the relationship holds even when the actual ppl scores are high (4/n)