Alexander Huth
alexanderhuth.bsky.social
Alexander Huth
@alexanderhuth.bsky.social
Interested in how & what the brain computes. Professor in Neuroscience & Statistics UC Berkeley
Always include some stimuli that are permissively licensed so they can be used as examples! E.g. we have a video stimulus set that's mostly Pixar short films, but also includes a segment from the Blender movie Sintel (en.wikipedia.org/wiki/Sintel), which is licensed CC-BY.
Sintel - Wikipedia
en.wikipedia.org
October 1, 2025 at 5:40 PM
Someone on the aggies subreddit posted the course description and it seems pretty thorough! www.reddit.com/r/aggies/s/C...
Wang_Lung_1921's comment on "Regardless of your stance on the politics behind the Welsh/LGBT situation…"
Explore this conversation and more from the aggies community
www.reddit.com
September 10, 2025 at 6:47 AM
New paper: Ask 35 simple questions about sentences in a story and use the answers to predict brain responses. Interpretable, compact, & surprisingly high performance in both fMRI and ECoG. 🧵 biorxiv.org/content/10.1...
August 18, 2025 at 6:34 PM
I'm posting this thread to highlight some things I thought cool, but if you're interested you should also check out what @rjantonello.bsky.social wrote: bsky.app/profile/rjan...
In our new paper, we explore how we can build encoding models that are both powerful and understandable. Our model uses an LLM to answer 35 questions about a sentence's content. The answers linearly contribute to our prediction of how the brain will respond to that sentence. 1/6
August 18, 2025 at 6:34 PM
Cortical weight maps were also reasonably correlated between ECoG and fMRI data, at least for the dimensions well-captured in the ECoG coverage.
August 18, 2025 at 6:34 PM
Finally, we tested whether the same interpretable embeddings could also be used to model ECoG data from Nima Mesgarani's lab. Despite the fact that our features are less well-localized in time than LLM embeddings, this still works quite well!
August 18, 2025 at 6:34 PM
To validate the maps we get from this model we also compared them to expectations derived from NeuroSynth and results from experiments targeting specific semantic categories, and also looked at inter-subject reliability. All quite successful.
August 18, 2025 at 6:34 PM
The model and experts were well-aligned, but there were some surprises, like "Does the input include technical or specialized terminology?" (32), which was much more important than expected.
August 18, 2025 at 6:34 PM
This method lets us quantitatively assess how much variance different theories explain about brain responses to natural language. So to figure out how well this aligns with what scientists think, we polled experts to see which questions/theories they thought would be important.
August 18, 2025 at 6:34 PM
"Does the input include dialogue?" (27) has high weights in a smattering of small regions in temporal cortex. And "Does the input contain a negation?" (35) has high weights in anterior temporal lobe and a few prefrontal areas. I think there's a lot of drilling-down we can do here.
August 18, 2025 at 6:34 PM
The fact that each dimension in the embedding thus corresponds to a specific question means that the encoding model weights are interpretable right out-of-the-box. "Does the input describe a visual experience?" has high weight all along the boundary of visual cortex, for example.
August 18, 2025 at 6:34 PM
But the wilder thing is how we get the embeddings: by just asking LLMs questions. Each theory is cast as a yes/no question. We then have GPT-4 answer each question about each 10-gram in our natural language dataset. We did this for ~600 theories/questions.
August 18, 2025 at 6:34 PM
And it works REALLY well! Prediction performance for encoding models is on a par with uninterpretable Llama3 embeddings! Even with just 35 dimensions!!! I find this fairly wild.
August 18, 2025 at 6:34 PM
So if you're interested in human memory for complex stimuli, please check the paper out! www.biorxiv.org/content/10.1...
Efficient uniform sampling explains non-uniform memory of narrative stories
Humans do not remember all experiences uniformly. We remember certain moments better than others, and central gist better than detail. Current theories focus exclusively on surprise to explain why som...
www.biorxiv.org
August 1, 2025 at 4:54 PM
This work was a really fun departure for me. Nothing data-driven (and no fMRI!), we just sat down and devised a theory, then tested it. It feels surprisingly good :D
August 1, 2025 at 4:54 PM
Our model also has interesting linguistic consequences. Speech tends to have uniform information density over time, but there are local variations. We argue that these variations (at least around event boundaries) are actually in service of more uniform _memory_ density.
August 1, 2025 at 4:54 PM
We also devised a new way to model and study gist with LLMs, which is to measure (or manipulate) the entropy of attention weights for specific "induction" heads within the LLM. Higher entropy attention weights more evenly sample information from the input, and lead to gist-like behavior.
August 1, 2025 at 4:54 PM