Michelle Lam
banner
mlam.bsky.social
Michelle Lam
@mlam.bsky.social
Stanford CS PhD student | hci, human-centered AI, social computing, responsible AI (+ dance, design, doodling!)
michelle123lam.github.io
A huge thank you to co-authors @fredhohman.bsky.social, @domoritz.de, @jeffreybigham.com, @kenholstein.bsky.social, and Mary Beth Kery! This work was done during my summer internship w/ Apple AIML, and I’m thankful to work with this wonderful team :)

arxiv.org/abs/2409.18203
#UIST25 talk: Wed 11am!
Policy Maps: Tools for Guiding the Unbounded Space of LLM Behaviors
AI policy sets boundaries on acceptable behavior for AI models, but this is challenging in the context of large language models (LLMs): how do you ensure coverage over a vast behavior space? We introd...
arxiv.org
September 29, 2025 at 3:54 PM
We can extend policy maps to enable Git-style collaboration and forking, aid live deliberation, and support longitudinal policy test suites & third-party audits. Policy maps can transform a nebulous space of model possibilities to an explicit specification of model behavior.
September 29, 2025 at 3:54 PM
With our system, LLM safety experts rapidly discovered policy gaps and crafted new policies around problematic model behavior (e.g., incorrectly assuming genders; repeating hurtful names in summaries; blocking physical safety threats that a user needs to be able to monitor).
September 29, 2025 at 3:54 PM
Given the unbounded space of LLM behaviors, developers need tools that concretize the subjective decisionmaking inherent to policy design. They should have a visual space to systematically explore, with explicit conceptual links between lofty principles and grounded examples.
September 29, 2025 at 3:54 PM
Our system creates linked map layers of cases, concepts, & policies: so an AI developer can author a policy that blocks model responses involving violence, visually notice a gap of physical threats that a user ought to be aware of, and test a revised policy to address this gap.
September 29, 2025 at 3:54 PM
We made updates to LLooM after the CHI publication to support local models (and non-OpenAI models)! More info here, though we haven't run evals across open-source models: stanfordhci.github.io/lloom/about/...
Custom Models | LLooM
Concept Induction: Analyzing Unstructured Text with High-Level Concepts
stanfordhci.github.io
January 17, 2025 at 11:46 PM
Qualitatively, I found that the BERTopic groupings were still rather large, so I anticipate the GPT labels would still be quite generic (as opposed to specific/targeted concepts).
January 17, 2025 at 5:18 PM
That's a good point! In the technical evaluations, we used GPT to automatically find matches between the methods (including a GPT-only condition), but it could have evened the playing field even more to generate GPT-style labels for BERTopic before the matching step.
January 17, 2025 at 5:18 PM
Thanks so much for sharing our work! :)
January 17, 2025 at 5:07 PM