Sander Dieleman
sedielem.bsky.social
Sander Dieleman
@sedielem.bsky.social
Blog: https://sander.ai/
🐦: https://x.com/sedielem
Research Scientist at Google DeepMind (WaveNet, Imagen 3, Veo, ...). I tweet about deep learning (research + software), music, generative models (personal account).
... also very honoured and grateful to see my blog linked in the video description! 🥹🙏🙇
July 26, 2025 at 9:59 PM
Everyone is welcome!
July 15, 2025 at 9:38 PM
Note also that getting this number slightly wrong isn't that big a deal. Even if you make it 100k instead of 10k, it's not going to change the granularity of the high frequencies that much because of the logarithmic frequency spacing.
June 24, 2025 at 11:39 PM
The frequencies are log-spaced, so historically, 10k was plenty to ensure that all positions can be uniquely distinguished. Nowadays of course sequences can be quite a bit longer.
June 24, 2025 at 11:39 PM
Yes! Also listen to this and contemplate the universe: grumusic.bandcamp.com/album/cosmog...
Cosmogenesis, by grumusic
8 track album
grumusic.bandcamp.com
January 28, 2025 at 11:53 PM
This is just a tiny fraction of what's available, check out the schedule for more: neurips.cc/virtual/2024...
NeurIPS 2024 Schedule
neurips.cc
January 22, 2025 at 9:04 PM
10. Last but not least (😎), here's my own workshop talk about multimodal iterative refinement: the methodological tension between language and perceptual modalities, autoregression and diffusion, and how to bring these together 🍸 neurips.cc/virtual/2024...
NeurIPS Multimodal Iterative RefinementNeurIPS 2024
neurips.cc
January 22, 2025 at 9:04 PM
9. A great overview of various strategies for merging multiple models together by Colin Raffel 🪿 neurips.cc/virtual/2024...
NeurIPS Colin RaffleNeurIPS 2024
neurips.cc
January 22, 2025 at 9:04 PM
8. Ishan Misra gives a nice overview of Meta's Movie Gen model 📽️ (I have some questions about the diffusion vs. flow matching comparison though😁) neurips.cc/virtual/2024...
NeurIPS Invited Talk 4 (Speker: Ishan Misra)NeurIPS 2024
neurips.cc
January 22, 2025 at 9:04 PM
7. More on test-time scaling from @tomgoldstein.bsky.social, using a different approach based on recurrence 🐚 neurips.cc/virtual/2024... (some interesting comments on the link with diffusion models in the questions at the end!)
NeurIPS Tom Goldstein: Can transformers solve harder problems than they were trained on? Scaling up test-time computation via recurrenceNeurIPS 2024
neurips.cc
January 22, 2025 at 9:04 PM
6. @polynoamial.bsky.social talks about scaling compute at inference time, and the trade-offs involved -- in language models, but also in other settings 🧮 neurips.cc/virtual/2024...
NeurIPS Invited Speaker: Noam Brown, OpenAINeurIPS 2024
neurips.cc
January 22, 2025 at 9:04 PM
5. Sparse autoencoders were in vogue well over a decade ago, back when I was doing my PhD. They've recently been revived in the context of mechanistic interpretability of LLMs 🔍 @neelnanda.bsky.social gives a nice overview: neurips.cc/virtual/2024...
NeurIPS Neel Nanda: Sparse Autoencoders - Assessing the evidenceNeurIPS 2024
neurips.cc
January 22, 2025 at 9:04 PM