And why does attention work so much better with multiple heads?
There might be a common answer to both of these questions.
Our #NeurIPS2025 Spotlight paper suggests that it can -- with the right training distribution.
🧵 A short thread:
Our #NeurIPS2025 Spotlight paper suggests that it can -- with the right training distribution.
🧵 A short thread:
Come checkout our Oral at #ICLR tomorrow (Apr 26th, poster at 10:00, Oral session 6C in the afternoon).
openreview.net/forum?id=V4K...
Come checkout our Oral at #ICLR tomorrow (Apr 26th, poster at 10:00, Oral session 6C in the afternoon).
openreview.net/forum?id=V4K...
Is mechanism modeling dead in the AI era?
ML models trained to predict neural activity fail to generalize to unseen opto perturbations. But mechanism modeling can solve that.
We say "perturbation testing" is the right way to evaluate mechanisms in data-constrained models
1/8
Is mechanism modeling dead in the AI era?
ML models trained to predict neural activity fail to generalize to unseen opto perturbations. But mechanism modeling can solve that.
We say "perturbation testing" is the right way to evaluate mechanisms in data-constrained models
1/8
The eighth of these, would you believe? We’ve got dark neurons, tiny monkeys, the most complete brain wiring diagram ever constructed, and much more…
Published on The Spike
Enjoy!
medium.com/the-spike/20...
The eighth of these, would you believe? We’ve got dark neurons, tiny monkeys, the most complete brain wiring diagram ever constructed, and much more…
Published on The Spike
Enjoy!
medium.com/the-spike/20...
This review aims to provide some intuition for and derivations of RL methods commonly used in systems neuroscience, ranging from TD learning through the SR to deep and distributional RL!
This review aims to provide some intuition for and derivations of RL methods commonly used in systems neuroscience, ranging from TD learning through the SR to deep and distributional RL!
Can neuroscience localizers uncover brain-like functional specializations in LLMs? 🧠🤖
Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks!
w/ @gretatuckute.bsky.social, @abosselut.bsky.social, @mschrimpf.bsky.social
🧵👇
Can neuroscience localizers uncover brain-like functional specializations in LLMs? 🧠🤖
Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks!
w/ @gretatuckute.bsky.social, @abosselut.bsky.social, @mschrimpf.bsky.social
🧵👇
The unts in ANN are actually not a terrible approximation of how real neurons work!
A tiny 🧵.
🧠📈 #NeuroAI #MLSky
I've seen people suggesting it's problematic, that neuroscientists won't like it, and so on.
But, I literally don't see why this is problematic...
Coming from the "giants" of AI.
Or maybe this was posted out of context? Please clarify.
I can't process this...
This will be the official account of the Eastern European Machine Learning (EEML) community.
Follow us for news regarding our summer schools, workshops, education/community initiatives, and more!
We study how task abstractions emerge in gated linear networks and how they support cognitive flexibility.
We study how task abstractions emerge in gated linear networks and how they support cognitive flexibility.
Reminds me of something Larry Abbott once said to me at a summer school:
Many physicists come into neuroscience assuming that the failure to find laws of the brain was just because biologists aren't clever enough. In fact, there are no laws.
🧠📈 🧪
Reminds me of something Larry Abbott once said to me at a summer school:
Many physicists come into neuroscience assuming that the failure to find laws of the brain was just because biologists aren't clever enough. In fact, there are no laws.
🧠📈 🧪
#neuroskyence
go.bsky.app/CAfmKQs
#neuroskyence
go.bsky.app/CAfmKQs
And why does attention work so much better with multiple heads?
There might be a common answer to both of these questions.
And why does attention work so much better with multiple heads?
There might be a common answer to both of these questions.