Michelle Lam
@mlam.bsky.social
900 followers 240 following 11 posts
Stanford CS PhD student | hci, human-centered AI, social computing, responsible AI (+ dance, design, doodling!) michelle123lam.github.io
Posts Media Videos Starter Packs
mlam.bsky.social
We can extend policy maps to enable Git-style collaboration and forking, aid live deliberation, and support longitudinal policy test suites & third-party audits. Policy maps can transform a nebulous space of model possibilities to an explicit specification of model behavior.
Broader usage scenarios inclue multi-stakeholder collaboration (live mode, git for policy, policy forks, participatory maps) and model evaluation + auditing (policy test suite, policy audits)
mlam.bsky.social
With our system, LLM safety experts rapidly discovered policy gaps and crafted new policies around problematic model behavior (e.g., incorrectly assuming genders; repeating hurtful names in summaries; blocking physical safety threats that a user needs to be able to monitor).
An evaluation with 12 LLM safety experts found it was much easier to identify policy gaps and author policies with the system compared to in their normal work.
mlam.bsky.social
Given the unbounded space of LLM behaviors, developers need tools that concretize the subjective decisionmaking inherent to policy design. They should have a visual space to systematically explore, with explicit conceptual links between lofty principles and grounded examples.
mlam.bsky.social
Our system creates linked map layers of cases, concepts, & policies: so an AI developer can author a policy that blocks model responses involving violence, visually notice a gap of physical threats that a user ought to be aware of, and test a revised policy to address this gap.
Policy maps chart LLM policy coverage over an unbounded space of model behaviors. Here, an AI practitioner is designing a policy for how an LLM should summarize violent text. Policy map abstractions (right) allow the policy designer to interactively author and test policies that govern a model’s behavior using if-then rules over concepts. The designer can create any desired concept by providing a simple text definition to capture cases of model behavior. Our Policy Projector tool (center) renders cases, concepts, and policies as visual map layers to aid iterative policy design.
mlam.bsky.social
LLM safety work often reasons over high-level policies (be helpful & polite), but must tackle on-the-ground cases (unsolicited money advice when stocks are mentioned). This can feel like driving on an unfamiliar road guided by a generic driver’s manual instead of a map. We introduce: Policy Maps 🗺️
Reposted by Michelle Lam
mariaa.bsky.social
Somehow only just became aware of LlooM, a toolkit that uses a combination of clustering and prompts to extract concepts and describe custom datasets — similar to a topic model. Looks nice, with lots of documentation and open colab notebooks!

Has anyone used it?

stanfordhci.github.io/lloom/about/
What is LLooM? | LLooM
Concept Induction: Analyzing Unstructured Text with High-Level Concepts
stanfordhci.github.io
mlam.bsky.social
We made updates to LLooM after the CHI publication to support local models (and non-OpenAI models)! More info here, though we haven't run evals across open-source models: stanfordhci.github.io/lloom/about/...
Custom Models | LLooM
Concept Induction: Analyzing Unstructured Text with High-Level Concepts
stanfordhci.github.io
mlam.bsky.social
Qualitatively, I found that the BERTopic groupings were still rather large, so I anticipate the GPT labels would still be quite generic (as opposed to specific/targeted concepts).
mlam.bsky.social
That's a good point! In the technical evaluations, we used GPT to automatically find matches between the methods (including a GPT-only condition), but it could have evened the playing field even more to generate GPT-style labels for BERTopic before the matching step.
mlam.bsky.social
Thanks so much for sharing our work! :)
mlam.bsky.social
We're excited to host a second iteration of the HEAL workshop! Join us at CHI 2025 :)

→ Deadline: Feb 17, more info at heal-workshop.github.io
liuyulu.bsky.social
Human-centered Evalulation and Auditing of Language models (HEAL) workshop is back for #CHI2025, with this year's special theme: “Mind the Context”! Come join us on this bridge between #HCI and #NLProc!

Workshop submission deadline: Feb 17 AoE
More info at heal-workshop.github.io.
The image includes a shortened call for participation that reads: 
"We welcome participants who work on topics related to supporting human-centered evaluation and auditing of language models. Topics of interest include, but not limited to:
- Empirical understanding of stakeholders' needs and goals of LLM evaluation and auditing
- Human-centered evaluation and auditing methods for LLMs
- Tools, processes, and guidelines for LLM evaluation and auditing
- Discussion of regulatory measures and public policies for LLM auditing
- Ethics in LLM evaluation and auditing

Special Theme: Mind the Context. We invite authors to engage with specific contexts in LLM evaluation and auditing. This theme could involve various topics: the usage contexts of LLMs, the context of the evaluation/auditing itself, and more! The term ''context'' is purposefully left open for interpretation!

The image also includes pictures of workshop organizers, who are: Yu Lu Liu, Wesley Hanwen Deng, Michelle S. Lam, Motahhare Eslami, Juho Kim, Q. Vera Liao, Wei Xu, Jekaterina Novikova, and Ziang Xiao.