Emma Harvey
@emmharv.bsky.social
990 followers 360 following 77 posts
PhD student @ Cornell info sci | Sociotechnical fairness & algorithm auditing | Previously MSR FATE, Penn | https://emmaharv.github.io/
Posts Media Videos Starter Packs
Pinned
emmharv.bsky.social
4. "The New York City Marathon", by me and 50,000 others @ all five boros this past weekend!
Me waving to the camera while running the NYC marathon
Reposted by Emma Harvey
queerinai.com
We are launching our Graduate School Application Financial Aid Program (www.queerinai.com/grad-app-aid) for 2025-2026. We’ll give up to $750 per person to LGBTQIA+ STEM scholars applying to graduate programs. Apply at openreview.net/group?id=Que.... 1/5
Grad App Aid — Queer in AI
www.queerinai.com
Reposted by Emma Harvey
tskuo.bsky.social
🌟 If you’re applying to CMU SCS PhD programs, and come from a background that would bring additional dimensions to the CMU community, our PhD students are here to help!

Apply to the Graduate Applicant Support Program by Oct 13 to receive feedback on your application materials:
Carnegie Mellon University School of Computer Science Graduate Application Support Program. Apply by October 13, 2025.
Reposted by Emma Harvey
Reposted by Emma Harvey
sandhaus.bsky.social
Join us for NYC Privacy Day 2025 at Cornell Tech, hosted by DLI @nissenbaum.bsky.social and SETS @mantzarlis.com.
We have a great selection of speakers and alongside
talks, we’ll feature student posters + demos.

🔗 Details, registration, and poster submission: dli.tech.cornell.edu/nyc-privacy-...
NYC Privacy Day 2025 | Cornell Tech
NYC Privacy Day hosted at Cornell Tech
dli.tech.cornell.edu
Reposted by Emma Harvey
travislloydphd.bsky.social
I'm at Seattle 4S! I'll be part of the "Risks of 'Social Model Collapse' in the Face of Scientific and Technological Advances" panel Friday morning, discussing online community governance of AI-generated content. Would love to meet others studying AI's impact on the info ecosystem!
#STS #4S
Reposted by Emma Harvey
paper-feed.bsky.social
**Please repost** If you're enjoying Paper Skygest -- our personalized feed of academic content on Bluesky -- we'd appreciate you reposting this! We’ve found that the most effective way for us to reach new users and communities is through users sharing it with their network
Reposted by Emma Harvey
avisokay.bsky.social
This is such a great paper and really helps to emphasize how data under specification in ML systems bias our understanding and decision making. Especially in inequitable resource scarce settings. Thanks for sharing @emmharv.bsky.social !
emmharv.bsky.social
👯 Allocation Multiplicity: Evaluating the Promises of the Rashomon Set by Jain et al. (incl. @kathleencreel.bsky.social) argues that allocation (vs. model) multiplicity should be seen as a pathway for reducing discrimination, homogenization, and arbitrariness in decision-making problems.
Screenshot of paper title and author list:

Allocation Multiplicity: Evaluating the Promises of the Rashomon Set
Shomik Jain, Margaret Wang, Kathleen Creel, Ashia Wilson
Reposted by Emma Harvey
rajiinio.bsky.social
Emma has such good research taste :)

Given the sheer scale of these events, its really helpful to see what caught people's eye at these conferences...
emmharv.bsky.social
After having such a great time at #CHI2025 and #FAccT2025, I wanted to share some of my favorite recent papers here!

I'll aim to post new ones throughout the summer and will tag all the authors I can find on Bsky. Please feel welcome to chime in with thoughts / paper recs / etc.!!

🧵⬇️:
emmharv.bsky.social
What's super cool to me about this paper is that it does a longitudinal analysis - so many audit studies stick to a single point in time, and this paper is a great demonstration that the data available to you at that time will inevitably impact your measurements.

🔗: dl.acm.org/doi/10.1145/...
Bias Delayed is Bias Denied? Assessing the Effect of Reporting Delays on Disparity Assessments | Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency
You will be notified whenever a record that you have chosen has been cited.
dl.acm.org
emmharv.bsky.social
The authors find that delays in reporting patient demographics are the norm (for >50% of patients, race is reported >60 days after other details like DOB). These delays obfuscate measurement of health outcomes and health disparities, and techniques like imputing race do not improve measurement.
emmharv.bsky.social
⏳ Bias Delayed is Bias Denied? Assessing the Effect of Reporting Delays on Disparity Assessments by @jennahgosciak.bsky.social and @aparnabee.bsky.social et al. (incl. @allisonkoe.bsky.social @marzyehghassemi.bsky.social) analyzes how missing demographic data impacts estimates of health disparities.
Screenshot of paper title and author list: 

Bias Delayed is Bias Denied? Assessing the Effect of Reporting Delays on Disparity Assessments 
Jennah Gosciak, Aparna Balagopalan, Derek Ouyang, Allison Koenecke, Marzyeh Ghassemi, Daniel E. Ho
emmharv.bsky.social
I love how this paper emphasizes that evaluation != accountability and recommends steps towards accountability: ensuring that tools are open, valid, and reliable; focusing on tools to support participatory methods; and ensuring auditors are protected from retaliation.

🔗: dl.acm.org/doi/full/10....
Towards AI Accountability Infrastructure: Gaps and Opportunities in AI Audit Tooling | Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems
You will be notified whenever a record that you have chosen has been cited.
dl.acm.org
emmharv.bsky.social
The authors analyze 435(!) tools, finding that most focus on evaluation – but tools for other aspects of AI audits, like harms discovery, communicating audit results, and advocating for change, are less common. Further, while many tools are freely available, auditors often struggle to use them.
emmharv.bsky.social
Towards AI Accountability Infrastructure: Gaps and Opportunities in AI Audit Tooling by @victorojewale.bsky.social @rbsteed.com @briana-v.bsky.social @abeba.bsky.social @rajiinio.bsky.social compares the landscape of AI audit tools (tools.auditing-ai.com) to the actual needs of AI auditors.
Screenshot of paper title and author list: 

Towards AI Accountability Infrastructure: Gaps and Opportunities in AI Audit Tooling
Victor Ojewale, Ryan Steed, Briana Vecchione, Abeba Birhane, Inioluwa Deborah Raji
emmharv.bsky.social
This work won a 🏆Best Paper Award🏆 at FAccT! I think it's a fantastic example of an external audit that not only identifies a problem but also provides concrete steps towards a solution.

🔗: dl.acm.org/doi/10.1145/...
External Evaluation of Discrimination Mitigation Efforts in Meta's Ad Delivery | Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency
dl.acm.org
emmharv.bsky.social
The authors show that the external review established by the settlement is insufficient to guarantee that Meta is actually reducing discrimination in ad delivery (as opposed to adversarially complying by showing the same ad repeatedly to one person or applying VRS only to small ad campaigns).
emmharv.bsky.social
📱 External Evaluation of Discrimination Mitigation Efforts in Meta's Ad Delivery by Imana et al. audits VRS (Meta’s process for reducing bias in ad delivery as part of a settlement with DOJ), and finds VRS reduces demographic differences in ad audience – but also reduces reach and increases cost.
Screenshot of paper title and author list: 

External Evaluation of Discrimination Mitigation Efforts in Meta’s Ad Delivery 
Basileal Imana, Zeyu Shen, John Heidemann, Aleksandra Korolova
emmharv.bsky.social
I thought this paper was really interesting, and I particularly appreciated the authors' point that models can make decisions that are "consistent" but still "arbitrary" if model selection is not done in a principled way!

🔗: dl.acm.org/doi/10.1145/...
dl.acm.org
emmharv.bsky.social
The authors propose that opportunity pluralism is most important in domains that involve normative or high-uncertainty decisions, or where decision-subject can choose among multiple decision-makers. Even in those domains, the authors argue that individual models should still be consistent.
emmharv.bsky.social
🎲 Consistently Arbitrary or Arbitrarily Consistent: Navigating the Tensions Between Homogenization and Multiplicity in Algorithmic Decision-Making by Gur-Arieh and Lee explores the competing desires for consistency in decision-making models and opportunity pluralism in decision-making ecosystems.
Screenshot of paper title and author list: 

Consistently Arbitrary or Arbitrarily Consistent: Navigating the Tensions Between Homogenization and Multiplicity in Algorithmic Decision-Making
Shira Gur-Arieh, Christina Lee
Reposted by Emma Harvey
allisonkoe.bsky.social
Check out our work at @ic2s2.bsky.social this afternoon during the Communication & Cooperation II session!
emmharv.bsky.social
I am so excited to be in 🇬🇷Athens🇬🇷 to present "A Framework for Auditing Chatbots for Dialect-Based Quality-of-Service Harms" by me, @kizilcec.bsky.social, and @allisonkoe.bsky.social, at #FAccT2025!!

🔗: arxiv.org/pdf/2506.04419
A screenshot of our paper's:

Title: A Framework for Auditing Chatbots for Dialect-Based Quality-of-Service Harms
Authors: Emma Harvey, Rene Kizilcec, Allison Koenecke
Abstract: Increasingly, individuals who engage in online activities are expected to interact with large language model (LLM)-based chatbots. Prior work has shown that LLMs can display dialect bias, which occurs when they produce harmful responses when prompted with text written in minoritized dialects. However, whether and how this bias propagates to systems built on top of LLMs, such as chatbots, is still unclear. We conduct a review of existing approaches for auditing LLMs for dialect bias and show that they cannot be straightforwardly adapted to audit LLM-based chatbots due to issues of substantive and ecological validity. To address this, we present a framework for auditing LLM-based chatbots for dialect bias by measuring the extent to which they produce quality-of-service harms, which occur when systems do not work equally well for different people. Our framework has three key characteristics that make it useful in practice. First, by leveraging dynamically generated instead of pre-existing text, our framework enables testing over any dialect, facilitates multi-turn conversations, and represents how users are likely to interact with chatbots in the real world. Second, by measuring quality-of-service harms, our framework aligns audit results with the real-world outcomes of chatbot use. Third, our framework requires only query access to an LLM-based chatbot, meaning that it can be leveraged equally effectively by internal auditors, external auditors, and even individual users in order to promote accountability. To demonstrate the efficacy of our framework, we conduct a case study audit of Amazon Rufus, a widely-used LLM-based chatbot in the customer service domain. Our results reveal that Rufus produces lower-quality responses to prompts written in minoritized English dialects.
Reposted by Emma Harvey
aclmeeting.bsky.social
🥳 🎉 ❤️ The ACL 2025 Proceedings are live on the ACL Anthology 🥰 !
We’re thrilled to pre-celebrate the incredible research 📚 ✨ that will be presented starting Monday next week in Vienna 🇦🇹 !
Start exploring 👉 aclanthology.org/events/acl-2...
#NLProc #ACL2025NLP #ACLAnthology
Annual Meeting of the Association for Computational Linguistics (2025) - ACL Anthology
pdf bibProceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)Wanxiang Che | Joyce Nabende | Ekaterina Shutova | Mohammad Taher Pilehvar
aclanthology.org
emmharv.bsky.social
I broke the thread 😅
emmharv.bsky.social
🚗 Not Even Nice Work If You Can Get It; A Longitudinal Study of Uber's Algorithmic Pay and Pricing by @rdbinns.bsky.social @jmlstein.bsky.social et al. (incl. @emax.bsky.social) audits Uber's pay practices, focusing on the shift to paying drivers a "dynamic" (opaque, unpredictable) share of fare.
Screenshot of paper title and author list:

Not Even Nice Work If You Can Get It; A Longitudinal Study of Uber's Algorithmic Pay and Pricing
Reuben Binns, Jake Stein, Siddhartha Datta, Max Van Kleek, Nigel Shadbolt
emmharv.bsky.social
My favorite part of this paper was the point that "synthetic data creates distance between individuals and the data...derived from [them]." Synthetic data is often considered privacy-preserving, but it can actually reduce opportunities for participation and redress!

🔗: dl.acm.org/doi/10.1145/...
Examining the Expanding Role of Synthetic Data Throughout the AI Development Pipeline | Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency
dl.acm.org