Pete Shaw
ptshaw.bsky.social
Pete Shaw
@ptshaw.bsky.social
Research Scientist at Google DeepMind. Mostly work on ML, NLP, and BioML. Based in Seattle.

http://ptshaw.com
We hope this work adds some conceptual clarity around how Kolmogorov complexity relates to neural networks, and provides a path towards identifying new complexity measures that enable greater compression and generalization.
October 1, 2025 at 2:11 PM
We prove that asymptotically optimal objectives exist for Transformers, building on a new demonstration of their computational universality. We also highlight potential challenges related to effectively optimizing such objectives.
October 1, 2025 at 2:11 PM
To address this question, we define the notion of asymptotically optimal description length objectives. We establish that a minimizer of such an objective achieves optimal compression, for any dataset, up to an additive constant, in the limit as model resource bounds increase.
October 1, 2025 at 2:11 PM
The Kolmogorov complexity of an object is the length of the shortest program that prints that object. Combining Kolmogorov complexity with the MDL principle provides an elegant foundation for formalizing Occam’s razor. But how can these ideas be applied to neural networks?
October 1, 2025 at 2:11 PM
Hi Marc, thanks for putting this together, mind adding me?
November 19, 2024 at 7:04 PM