🎓 https://scholar.google.com/citations?hl=en&user=k5eR8_oAAAAJ
We found only feedback+dropout constrained model activity on low dimensional manifolds!
We found only feedback+dropout constrained model activity on low dimensional manifolds!
Crucially, only models with both top-down feedback and dropout (dark blue 👇) outperformed all other models on sensory robustness
Crucially, only models with both top-down feedback and dropout (dark blue 👇) outperformed all other models on sensory robustness
🔹 Feedback connections (present / absent)
🔵 Dropout (present / absent during training)
Models had 3 layers, 10 time steps, with 1 long-range feedback connection from last to first layer
🔹 Feedback connections (present / absent)
🔵 Dropout (present / absent during training)
Models had 3 layers, 10 time steps, with 1 long-range feedback connection from last to first layer
➡️ recurrent (lateral), reinjecting layer’s previous hidden state
↘️ feedback, projecting from higher to lower layers
leaving the individual roles of each pathway unclear
➡️ recurrent (lateral), reinjecting layer’s previous hidden state
↘️ feedback, projecting from higher to lower layers
leaving the individual roles of each pathway unclear
We found ConvRNN with top-down feedback exhibiting OOD robustness only when trained with dropout, revealing a dual mechanism for robust sensory coding
with @marco-d.bsky.social, Karl Friston, Giovanni Pezzulo & @siegellab.bsky.social
🧵👇
We found ConvRNN with top-down feedback exhibiting OOD robustness only when trained with dropout, revealing a dual mechanism for robust sensory coding
with @marco-d.bsky.social, Karl Friston, Giovanni Pezzulo & @siegellab.bsky.social
🧵👇
👇 model representations for natural, synth, and defeat. stimuli
👇 model representations for natural, synth, and defeat. stimuli
👇 human responses in the natural, synth and “out-of-distribution” (defeat.) case, where the response pattern is not random
👇 human responses in the natural, synth and “out-of-distribution” (defeat.) case, where the response pattern is not random
Synthesized (ITD & ILD) stimuli had ✅ITD, ✅ILD but not ❌third features
Defeaturized (ITD & ILD) stimuli had ✅third features but not ❌ITD, ❌ILD
Synthesized (ITD & ILD) stimuli had ✅ITD, ✅ILD but not ❌third features
Defeaturized (ITD & ILD) stimuli had ✅third features but not ❌ITD, ❌ILD
Synthesized stimuli were generated by adding these cues to monaural sounds
Defeaturized stimuli were generated by subtracting these cues from the recorded sounds
Synthesized stimuli were generated by adding these cues to monaural sounds
Defeaturized stimuli were generated by subtracting these cues from the recorded sounds
We compared humans and deep neural networks on sound localization 👂📍
Humans robustly localized OOD sounds even without primary interaural cues (ITD & ILD)
Models localized well only in-training distribution sounds, failing on OOD regime
Link & full story 🧵👇
We compared humans and deep neural networks on sound localization 👂📍
Humans robustly localized OOD sounds even without primary interaural cues (ITD & ILD)
Models localized well only in-training distribution sounds, failing on OOD regime
Link & full story 🧵👇
it's always been and always will be about OOD generalization, probably the unifying problem that defines if a system is intelligent or not
The key question is, what inductive bias do you use to solve this task?
👇
arxiv.org/abs/2507.06952
it's always been and always will be about OOD generalization, probably the unifying problem that defines if a system is intelligent or not
The key question is, what inductive bias do you use to solve this task?
👇
arxiv.org/abs/2507.06952
"We argue here that the most exciting of these is the use of AI models as cognitive models
[...]
Such cognitive models constitute a substantial advance that can inform theories of human intelligence"
#neuroAI
"We argue here that the most exciting of these is the use of AI models as cognitive models
[...]
Such cognitive models constitute a substantial advance that can inform theories of human intelligence"
#neuroAI