Lucy Li
banner
lucy3.bsky.social
Lucy Li
@lucy3.bsky.social
Postdoc at UW NLP 🏔️. #NLProc, computational social science, cultural analytics, responsible AI. she/her. Previously at Berkeley, Ai2, MSR, Stanford. Incoming assistant prof at Wisconsin CS. lucy3.github.io/prospective-students.html
Reposted by Lucy Li
🎉 Congratulations to all #EMNLP2025 award winners 🎉

Starting with the ✨Best Paper award ✨:

"Infini-gram mini: Exact n-gram Search at the Internet Scale with FM-Index"
by Hao Xu, Jiacheng Liu, Yejin Choi, Noah A. Smith, and Hannaneh Hajishirzi
aclanthology.org/2025.emnlp-m...

1/n
November 7, 2025 at 10:29 PM
It's the season for PhD apps!! 🥧 🦃 ☃️ ❄️

Apply to Wisconsin CS to research
- Societal impact of AI
- NLP ←→ CSS and cultural analytics
- Computational sociolinguistics
- Human-AI interaction
- Culturally competent and inclusive NLP
with me!

lucy3.github.io/prospective-...
November 11, 2025 at 10:32 PM
Reposted by Lucy Li
COLM is going to San Francisco for 2026!

🗓️Dates: October 6-9, 2026
🏨Venue: Hilton San Francisco Union Square

Website and CFPs for papers and workshops coming up soon!
November 11, 2025 at 7:30 PM
Reposted by Lucy Li
🌉 #EMNLP2026 will be October 24-29th in Budapest! 🌉

Thanks all for a great conference, and see you at the next one!
November 7, 2025 at 10:41 PM
Reposted by Lucy Li
I'm leading a session with PhD students on developing elevator pitches for research. Does anyone have suggested readings, websites, or just general advice on the topic?

No one ever taught me how to do this and honestly I struggle with this skill. All suggestions welcome!
November 11, 2025 at 10:13 PM
Not to brag too much or anything but after emnlp I got brunch with a llama in Shanghai
November 10, 2025 at 7:03 AM
If you don't already live in China and you're going to EMNLP, download WeChat / Alipay if you want to buy anything in the city and setup an eSIM on your phone before you leave your home country
October 29, 2025 at 10:19 PM
Reposted by Lucy Li
Cornell (NYC and Ithaca) is recruiting AI postdocs, apply by Nov 20, 2025! If you're interested in working with me on technical approaches to responsible AI (e.g., personalization, fairness), please email me.

academicjobsonline.org/ajo/jobs/30971
Cornell University, Empire AI Fellows Program
Job #AJO30971, Postdoctoral Fellow, Empire AI Fellows Program, Cornell University, New York, New York, US
academicjobsonline.org
October 28, 2025 at 6:19 PM
Reposted by Lucy Li
Ok y'all I'm throwing out a hot take on LLMs in Toronto in January:

"We’re Talking About the Wrong Error: Why Variance Matters More than Bias in AI"

Enough of the bias talk. LLMs are a completely different beast and our old frameworks are no longer useful.

datasciences.utoronto.ca/dsi-home/dat...
Data Sciences Speaker Series - DSI
The Data Sciences Speaker Series is a collaboration of data science programs at U of T. Seminars are held on the third Monday of each month.
datasciences.utoronto.ca
October 29, 2025 at 7:25 PM
tfw I have a problem but then I find some nice linguistics paper from pre-2000s that not only addresses the problem but also provides a very nice articulation of the problem
October 27, 2025 at 5:35 PM
Reposted by Lucy Li
❗REMINDER: The I School is hiring❗We are looking for an Assistant Professor of Information who can address key questions about metadata, information systems, and more. #AcademicSky

The final review date is Nov. 1. We look forward to reading your applications! https://aprecruit.berkeley.edu/JPF05014
Assistant Professor - Information - School of Information
University of California, Berkeley is hiring. Apply now!
aprecruit.berkeley.edu
October 27, 2025 at 4:57 PM
Reposted by Lucy Li
We're excited to host a virtual workshop where graduate students can present work-in-progress that engages with data from the Post45 Data Collective in some way.

Editors from the P45DC, Public Books, Post45, and the 19th Century DC will be there to offer feedback.

Proposals are due December 1!
October 27, 2025 at 5:17 PM
Reposted by Lucy Li
LLMs are often used for text annotation, especially in social science. In some cases, this involves placing text items on a scale: eg, 1 for liberal and 9 for conservative

There are a few ways to accomplish this task. Which work best? Our new EMNLP paper has some answers🧵
arxiv.org/pdf/2507.00828
October 27, 2025 at 2:59 PM
Reposted by Lucy Li
Aaaaarrrggh
October 26, 2025 at 4:17 PM
Reposted by Lucy Li
AI is already at work in American newsrooms.

We examine 186k articles published this summer and find that ~9% are either fully or partially AI-generated, usually without readers having any idea.

Here's what we learned about how AI is influencing local and national journalism:
October 22, 2025 at 3:24 PM
Strangest autocomplete-induced error: I had imported 🐢 and not 🐼
October 22, 2025 at 8:41 PM
Reposted by Lucy Li
Sorry to have to leave AIES early. I have to be back in Atlanta to run the Georgia Tech Summit on Responsible Computing, AI, and Society rcais.github.io
2025 Summit on Responsible Computing, AI, and Society
2025 Summit on Responsible Computing, AI, and Society
rcais.github.io
October 22, 2025 at 7:05 AM
Reposted by Lucy Li
As of June 2025, 66% of Americans have never used ChatGPT.

Our new position paper, Attention to Non-Adopters, explores why this matters: AI research is being shaped around adopters—leaving non-adopters’ needs, and key LLM research opportunities, behind.

arxiv.org/abs/2510.15951
October 21, 2025 at 5:12 PM
Reposted by Lucy Li
We discovered that language models leave a natural "signature" on their API outputs that's extremely hard to fake. Here's how it works 🔍

📄 arxiv.org/abs/2510.14086 1/
Every Language Model Has a Forgery-Resistant Signature
The ubiquity of closed-weight language models with public-facing APIs has generated interest in forensic methods, both for extracting hidden model details (e.g., parameters) and for identifying...
arxiv.org
October 17, 2025 at 5:59 PM
Reposted by Lucy Li
wikipedia's data shows that AI is siphoning traffic away from the site, which is a danger to its sustainability. ironically Wikipedia is more important than ever to users who want reliable information instead of slop, and to AI companies that need it for training data www.404media.co/wikipedia-sa...
Wikipedia Says AI Is Causing a Dangerous Decline in Human Visitors
“With fewer visits to Wikipedia, fewer volunteers may grow and enrich the content, and fewer individual donors may support this work.”
www.404media.co
October 17, 2025 at 1:15 AM
Reposted by Lucy Li
𝑵𝒆𝒘 𝒃𝒍𝒐𝒈𝒑𝒐𝒔𝒕! A rundown of some cool papers I got to chat about at #COLM2025 and some scattered thoughts

saxon.me/blog/2025/co...
COLM 2025: 9 cool papers and some thoughts
Reflections on the 2025 COLM conference, and a discussion of 9 cool COLM papers on benchmarking and eval, personas, and improving models for better long-context performance and consistency.
saxon.me
October 17, 2025 at 5:24 AM
I reward myself with a pop tart every time i say no to something
I also get a pop tart when I want a pop tart
I'm going through sooo many pop tarts
October 17, 2025 at 5:18 PM
would be so excited to see cognitive scientists and llm researchers working together on user modeling research
October 15, 2025 at 12:52 AM
Reposted by Lucy Li
𝐃𝐨 𝐲𝐨𝐮 𝐫𝐞𝐚𝐥𝐥𝐲 𝐰𝐚𝐧𝐭 𝐭𝐨 𝐬𝐞𝐞 𝐰𝐡𝐚𝐭 𝐦𝐮𝐥𝐭𝐢𝐥𝐢𝐧𝐠𝐮𝐚𝐥 𝐞𝐟𝐟𝐨𝐫𝐭 𝐥𝐨𝐨𝐤𝐬 𝐥𝐢𝐤𝐞? 🇨🇳🇮🇩🇸🇪

Here’s the proof! 𝐁𝐚𝐛𝐲𝐁𝐚𝐛𝐞𝐥𝐋𝐌 is the first Multilingual Benchmark of Developmentally Plausible Training Data available for 45 languages to the NLP community 🎉

arxiv.org/abs/2510.10159
October 14, 2025 at 5:01 PM
Reposted by Lucy Li
I am on the job market this year! My research advances methods for reliable machine learning from real-world data, with a focus on healthcare. Happy to chat if this is of interest to you or your department/team.
October 14, 2025 at 3:45 PM