Yeqing Wang
yeqingwang.bsky.social
Yeqing Wang
@yeqingwang.bsky.social
Reposted by Yeqing Wang
Why don’t neural networks learn all at once, but instead progress from simple to complex solutions? And what does “simple” even mean across different neural network architectures?

Sharing our new paper @iclr_conf led by Yedi Zhang with Peter Latham

arxiv.org/abs/2512.20607
February 3, 2026 at 4:19 PM
Reposted by Yeqing Wang
Congrats to Ella for her new paper! She asked a really interesting question about how the brain represents uncertainty during hidden state inference, and in a lovely crossover with theoretical work, she shows that in mice, acetylcholine dynamics play a crucial role. www.biorxiv.org/content/10.1...
Acetylcholine reflects uncertainty during hidden state inference
To act adaptively, animals must infer features of the environment that cannot be observed directly, such as which option is currently rewarding, or which context they are in. These internal estimates,...
www.biorxiv.org
November 14, 2025 at 9:57 AM
Reposted by Yeqing Wang
We watched mice follow scent trails drawn on an “endless” treadmill. By manipulating trail geometry/statistics, perturbing mouse nose & brain, and modeling behavior with a Bayesian framework, we show that mice use predictive (rather than reactive) strategies.
September 1, 2025 at 7:07 PM
Reposted by Yeqing Wang
How does in-context learning emerge in attention models during gradient descent training?

Sharing our new Spotlight paper @icmlconf.bsky.social: Training Dynamics of In-Context Learning in Linear Attention
arxiv.org/abs/2501.16265

Led by Yedi Zhang with @aaditya6284.bsky.social and Peter Latham
June 4, 2025 at 11:22 AM