Vera Neplenbroek
veraneplenbroek.bsky.social
Vera Neplenbroek
@veraneplenbroek.bsky.social
PhD student at ILLC / University of Amsterdam, interested in safety, bias, and stereotypes in conversational and generative AI #NLProc
https://veranep.github.io/
Next week, I'll be at #EMNLP presenting our work "Reading Between the Prompts: How Stereotypes Shape LLM's Implicit Personalization" 🎉

📍 Ethics, Bias, and Fairness (Poster Session 2)
📅 Wed, November 5, 11:00-12:30 - Hall C
📖 Check the paper! arxiv.org/abs/2505.16467

See you in Suzhou! 👋
October 31, 2025 at 7:56 PM
[3/4] Our findings reveal that LLMs infer demographic info based on stereotypical signals, sometimes even when the user explicitly identifies with a different demographic group. We mitigate this by intervening on the model’s internal representations using a trained linear probe.
May 27, 2025 at 10:41 AM
Do LLMs assume demographic information based on stereotypes?

We (@arianna-bis.bsky.social, Raquel Fernández and I) answered this question in our new paper: "Reading Between the Prompts: How Stereotypes Shape LLM's Implicit Personalization".

🧵

arxiv.org/abs/2505.16467
May 27, 2025 at 10:41 AM
[3/4] Our findings reveal that LLMs do infer demographic info based on stereotypical signals, sometimes even when the user explicitly identifies with a different demographic group. We mitigate this by intervening on the model’s
internal representations using a trained linear
probe.
May 27, 2025 at 10:37 AM
Happy to share that "LLMs instead of Human Judges? A Large Scale Empirical Study across 20 NLP Evaluation Tasks" arxiv.org/abs/2406.18403 got accepted to ACL Main! #ACL2025 🎉
May 21, 2025 at 2:38 PM
Excited to share that "Cross-Lingual Transfer of Debiasing and Detoxification in Multilingual LLMs: An Extensive Investigation" arxiv.org/abs/2412.14050 got accepted to ACL Findings! 🎉 #ACL2025 Big thanks to my supervisors Raquel Fernández and @arianna-bis.bsky.social for their guidance and support!
May 21, 2025 at 2:35 PM