Anima Anandkumar
banner
anima-anandkumar.bsky.social
Anima Anandkumar
@anima-anandkumar.bsky.social
AI Pioneer, AI+Science, Professor at Caltech, Former Senior Director of AI at NVIDIA, Former Principal Scientist at AWS AI.
End-to-end learning can use both approximate and accurate training data, if the model can learn how to mix them correctly. It turns out that Neural Operators offer a perfect solution when such multi-fidelity and multi-resolution data is available, and can learn with high data efficiency.
September 2, 2025 at 12:41 AM
Our latest paper surprisingly shows that it is not the case! End to end also requires less training data compared to methods that keep existing numerical solvers and augment with AI. Where do savings come from? The approach that augments AI relies only on fully accurate expensive training data.
September 2, 2025 at 12:40 AM
We have seen end-to-end approach win in areas like weather forecasting. It is significantly better for speed: 1000-million x faster than numerical simulations in many areas such as fluid dynamics, plasma physics etc. But a big argument against it is the need for expensive training data.
September 2, 2025 at 12:38 AM
Popular prescription is to augment AI into existing workflows rather than replace them, e.g., keep the approximate numerical solver for simulations, and use AI only to correct its errors in every time step. Other extreme is to completely discard the existing workflow and replace it fully with AI.
September 2, 2025 at 12:37 AM
Reposted by Anima Anandkumar
Thanks to my co-authors David Pitt, Robert Joseph George, Jiawwei Zhao, Cheng Luo, Yuandong Tian, Jean Kossaifi, @anima-anandkumar.bsky.social, and @caltech.edu for hosting me this spring!
Paper: arxiv.org/abs/2501.02379
Code: github.com/neuraloperat...
TensorGRaD: Tensor Gradient Robust Decomposition for Memory-Efficient Neural Operator Training
Scientific problems require resolving multi-scale phenomena across different resolutions and learning solution operators in infinite-dimensional function spaces. Neural operators provide a powerful fr...
arxiv.org
June 3, 2025 at 3:17 AM