Brenden Lake
brendenlake.bsky.social
Brenden Lake
@brendenlake.bsky.social
Associate Professor of Computer Science and Psychology @ Princeton. Posts are my views only. https://www.cs.princeton.edu/~bl8144/
The best write up on the state of AI in a while, from James Somers, with input from my Princeton colleagues Ken Norman, Uri Hasson, Jon Cohen, and many others. Coding with LLMs was a striking moment for me too www.newyorker.com/magazine/202...
The Case That A.I. Is Thinking
ChatGPT does not have an inner life. Yet it seems to know what it’s talking about.
www.newyorker.com
November 4, 2025 at 3:10 PM
Checking out the Princeton trails on our lab retreat
November 3, 2025 at 5:44 PM
There are still open desks in our new Human & Machine Intelligence lab at Princeton. Express your interest in joining us: lake-lab.github.io/apply/
October 22, 2025 at 1:35 PM
Today in Nature Machine Intelligence, Kazuki Irie & I discuss 4 classic challenges for neural nets — systematic generalization, catastrophic forgetting, few-shot learning, & reasoning. We argue there is a unifying fix: the right incentives & practice. rdcu.be/eLRmg
October 20, 2025 at 1:18 PM
Our new lab for Human & Machine Intelligence is officially open at Princeton University!

Consider applying for a PhD or Postdoc position, either through Computer Science or Psychology. You can register interest on our new website lake-lab.github.io (1/2)
September 8, 2025 at 1:59 PM
Reposted by Brenden Lake
For much more, see the paper! arxiv.org/abs/2508.05776

By Tom Griffiths, Brenden Lake, Tom McCoy, Ellie Pavlick, and Taylor Webb (@cocoscilab.bsky.social‬, @brendenlake.bsky.social‬, ‪@rtommccoy.bsky.social‬, Ellie Pavlick, @taylorwwebb.bsky.social‬)

9/9
Whither symbols in the era of advanced neural networks?
Some of the strongest evidence that human minds should be thought about in terms of symbolic systems has been the way they combine ideas, produce novelty, and learn quickly. We argue that modern neura...
arxiv.org
August 15, 2025 at 4:30 PM
Getting the lab + alums together at CogSci!
August 1, 2025 at 4:44 PM
I'm joining Princeton University as an Associate Professor of Computer Science and Psychology this fall! Princeton is ambitiously investing in AI and Natural & Artificial Minds, and I'm excited for my lab to contribute. Recruiting postdocs and Ph.D. students in CS and Psychology — join us!
June 12, 2025 at 2:29 PM
Reposted by Brenden Lake
Fantastic new work by @johnchen6.bsky.social (with @brendenlake.bsky.social and me trying not to cause too much trouble).

We study systematic generalization in a safety setting and find LLMs struggle to consistently respond safely when we vary how we ask naive questions. More analyses in the paper!
Do LLMs show systematic generalization of safety facts to novel scenarios?

Introducing our work SAGE-Eval, a benchmark consisting of 100+ safety facts and 10k+ scenarios to test this!

- Claude-3.7-Sonnet passes only 57% of facts evaluated
- o1 and o3-mini passed <45%! 🧵
May 30, 2025 at 5:32 PM
Failures of systematic generalization in LLMs can lead to real-world safety issues.

New paper by @johnchen6.bsky.social and @guydav.bsky.social, arxiv.org/abs/2505.21828
Do LLMs show systematic generalization of safety facts to novel scenarios?

Introducing our work SAGE-Eval, a benchmark consisting of 100+ safety facts and 10k+ scenarios to test this!

- Claude-3.7-Sonnet passes only 57% of facts evaluated
- o1 and o3-mini passed <45%! 🧵
May 29, 2025 at 7:17 PM
Before LLMs, neural nets were task-specific (while humans were task-general). Shockingly, LLMs changed that. How do LLMs represent a task, and do different prompts lead to the same task rep.? Love this by @guydav.bsky.social, and the function vectors of @ericwtodd.bsky.social @davidbau.bsky.social
New preprint alert! We often prompt ICL tasks using either demonstrations or instructions. How much does the form of the prompt matter to the task representation formed by a language model? Stick around to find out 1/N
May 23, 2025 at 8:24 PM
Reposted by Brenden Lake
🤔 Interested in models of social interaction and computational psychiatry?

🤗 If so, @shawnrhoadsphd.bsky.social and I are seeking a highly motivated and talented postdoc to work on these topics!

Please share widely!

apply.interfolio.com/165809
April 4, 2025 at 2:52 PM
Reposted by Brenden Lake
Despite the world being on fire, I can't help but be thrilled to announce that I'll be starting as an Assistant Professor in the Cognitive Science Program at Dartmouth in Fall '26. I'll be recruiting grad students this upcoming cycle—get in touch if you're interested!
May 7, 2025 at 10:08 PM
Reposted by Brenden Lake
New work in Nature Machine Intelligence by @guydav.bsky.social, @brendenlake.bsky.social, Todd Gureckis, Graham Todd, and Julian Togelius models how humans develop goals—research that could help bridge the gap between human intentions and AI systems.

nyudatascience.medium.com/what-is-a-go...
What is a Goal? Advancing Machine Agency While Understanding Human Goal Creation
New CDS research models how people understand and formulate goals — knowledge that could improve AI alignment
nyudatascience.medium.com
April 10, 2025 at 7:43 PM
I snuck a moment with my son Logan (2.5), ever the creative goal generator, into Fig. 1: "Papa, I made a Truck Carrier Truck!"
How do people compose existing concepts to create new goals? Can models generate and understand goals too?
nature.com/articles/s4225
February 21, 2025 at 6:41 PM
@solimlegris.bsky.social and Wai Keen Vong estimated that average human performance on ARC is about 64%(public eval set). Thus, o3 is clearly better than the average crowd worker tested. Note that almost all tasks were solvable by at least one person who tried it on MTurk. arxiv.org/abs/2409.01374
December 20, 2024 at 7:14 PM
I'm new here. I heard bluesky is like science Twitter back in the day, and there are fewer posts from Elon Musk. Did I come to the right place?
December 19, 2024 at 6:26 PM