Studying in-context learning and reasoning in humans and machines
Prev. @UofT CS & Psych
Our work explains this & *predicts Transformer behavior throughout training* without its weights! 🧵
1/
Our Temporal Feature Analyzer discovers contextual features in LLMs, that detect event boundaries, parse complex grammar, and represent ICL patterns.
Our Temporal Feature Analyzer discovers contextual features in LLMs, that detect event boundaries, parse complex grammar, and represent ICL patterns.
"Chain of Time: In-Context Physical Simulation with Image Generation Models"
(by Wang, Bigelow, Li, and me)
arxiv.org/abs/2511.00110
We examine how people figure out what happened by combining visual and auditory evidence through mental simulation.
Paper: osf.io/preprints/ps...
Code: github.com/cicl-stanfor...
We examine how people figure out what happened by combining visual and auditory evidence through mental simulation.
Paper: osf.io/preprints/ps...
Code: github.com/cicl-stanfor...
Humans are capable of sophisticated theory of mind, but when do we use it?
We formalize & document a new cognitive shortcut: belief neglect — inferring others' preferences, as if their beliefs are correct🧵
Humans are capable of sophisticated theory of mind, but when do we use it?
We formalize & document a new cognitive shortcut: belief neglect — inferring others' preferences, as if their beliefs are correct🧵
🧠 Looking for insight on applying to PhD programs in psychology?
✨ Apply by Sep 25th to Stanford Psychology's 9th annual Paths to a Psychology PhD info-session/workshop to have all of your questions answered!
📝 Application: tinyurl.com/pathstophd2025
🧠 Looking for insight on applying to PhD programs in psychology?
✨ Apply by Sep 25th to Stanford Psychology's 9th annual Paths to a Psychology PhD info-session/workshop to have all of your questions answered!
📝 Application: tinyurl.com/pathstophd2025
sites.google.com/view/infocog...
sites.google.com/view/infocog...
How can we interpret the algorithms and representations underlying complex behavior in deep learning models?
🌐 coginterp.github.io/neurips2025/
1/4
How can we interpret the algorithms and representations underlying complex behavior in deep learning models?
🌐 coginterp.github.io/neurips2025/
1/4
How can we interpret the algorithms and representations underlying complex behavior in deep learning models?
🌐 coginterp.github.io/neurips2025/
1/4
Our work explains this & *predicts Transformer behavior throughout training* without its weights! 🧵
1/
Our work explains this & *predicts Transformer behavior throughout training* without its weights! 🧵
1/
Our work explains this & *predicts Transformer behavior throughout training* without its weights! 🧵
1/
Once a cornerstone for studying human reasoning, the think-aloud method declined in popularity as manual coding limited its scale. We introduce a method to automate analysis of verbal reports and scale think-aloud studies. (1/8)🧵
Once a cornerstone for studying human reasoning, the think-aloud method declined in popularity as manual coding limited its scale. We introduce a method to automate analysis of verbal reports and scale think-aloud studies. (1/8)🧵