Video: youtube.com/live/DXQ7FZA...
Big thanks to the jury @dlarlus.bsky.social @ptrkprz.bsky.social @gtolias.bsky.social A. Efros & T. Karras
Video: youtube.com/live/DXQ7FZA...
Big thanks to the jury @dlarlus.bsky.social @ptrkprz.bsky.social @gtolias.bsky.social A. Efros & T. Karras
I was so in awe of the presentation that I even forgot to take pictures 😅
I was so in awe of the presentation that I even forgot to take pictures 😅
We would like to inform authors that the OR submission system requires the submission of a compute reporting form along with any updates to the full paper. We have identified this as a system-related issue.
1/2
We would like to inform authors that the OR submission system requires the submission of a compute reporting form along with any updates to the full paper. We have identified this as a system-related issue.
1/2
To me, the only solution to the credit assignment problem is obvious: stop believing a single person is responsible for every big discovery. It's an artifact of our monkey brain requiring a face for storage, not the reality of how knowledge progresses.
To me, the only solution to the credit assignment problem is obvious: stop believing a single person is responsible for every big discovery. It's an artifact of our monkey brain requiring a face for storage, not the reality of how knowledge progresses.
- 19x faster convergence ⚡
- 370x less FLOPS than FLUX-dev 📉
No more post-training alignment!
We integrate human alignment right from the start, during pretraining!
Results:
✨ 19x faster convergence ⚡
✨ 370x less compute 💻
🔗 Explore the project: nicolas-dufour.github.io/miro/
No more post-training alignment!
We integrate human alignment right from the start, during pretraining!
Results:
✨ 19x faster convergence ⚡
✨ 370x less compute 💻
🔗 Explore the project: nicolas-dufour.github.io/miro/
Check it out 👌
Check it out 👌
Read the thread for all the great details.
The main conclusion I draw from this work is that better pretraining, in particular by conditioning on better data, allows us to train SOTA models at a fraction of the cost.
- 19x faster convergence ⚡
- 370x less FLOPS than FLUX-dev 📉
Read the thread for all the great details.
The main conclusion I draw from this work is that better pretraining, in particular by conditioning on better data, allows us to train SOTA models at a fraction of the cost.
- 19x faster convergence ⚡
- 370x less FLOPS than FLUX-dev 📉
- 19x faster convergence ⚡
- 370x less FLOPS than FLUX-dev 📉
It might not be the easiest intro to diffusion models, but this monograph is an amazing deep dive into the math behind them and all the nuances
It might not be the easiest intro to diffusion models, but this monograph is an amazing deep dive into the math behind them and all the nuances
Who are those guys and in what society do they live?
This is exactly why I think the concept of AGI is meaningless.
Who are those guys and in what society do they live?
This is exactly why I think the concept of AGI is meaningless.
arxiv.org/abs/2505.22109
🏆 GRALE 🏆 can encode and decode graphs into and from a shared Euclidean space.
Training such a model should require solving the graph matching problem but...
As a PyTorch Ambassador, I would like to write an article to introduce open-sourced ICCV 2025 works (including workshops and demos) for promoting open-source/science + PyTorch
If interested, share your work via the form in my reply 👇
As a PyTorch Ambassador, I would like to write an article to introduce open-sourced ICCV 2025 works (including workshops and demos) for promoting open-source/science + PyTorch
If interested, share your work via the form in my reply 👇
Today I want to share two new works on this topic:
Eliciting higher alignment: arxiv.org/abs/2510.02425
Unpaired learning of unified reps: arxiv.org/abs/2510.08492
1/9
Today I want to share two new works on this topic:
Eliciting higher alignment: arxiv.org/abs/2510.02425
Unpaired learning of unified reps: arxiv.org/abs/2510.08492
1/9
Do you think there would be people in for that? Do you think it would make for a nice competition?
Do you think there would be people in for that? Do you think it would make for a nice competition?
TL;DR: train a text2image model from scratch on ImageNet only and beat SDXL.
Paper, code, data available! Reproducible science FTW!
🧵👇
📜 arxiv.org/abs/2502.21318
💻 github.com/lucasdegeorg...
💽 huggingface.co/arijitghosh/...
TL;DR: train a text2image model from scratch on ImageNet only and beat SDXL.
Paper, code, data available! Reproducible science FTW!
🧵👇
📜 arxiv.org/abs/2502.21318
💻 github.com/lucasdegeorg...
💽 huggingface.co/arijitghosh/...