Lucas Degeorge
@lucasdegeorge.bsky.social
PhD student at École Polytechnique (Vista) and École des Ponts (IMAGINE)
Working on conditional diffusion models
Working on conditional diffusion models
Check out our new work: MIRO
No more post-training alignment!
We integrate human alignment right from the start, during pretraining!
Results:
✨ 19x faster convergence ⚡
✨ 370x less compute 💻
🔗 Explore the project: nicolas-dufour.github.io/miro/
No more post-training alignment!
We integrate human alignment right from the start, during pretraining!
Results:
✨ 19x faster convergence ⚡
✨ 370x less compute 💻
🔗 Explore the project: nicolas-dufour.github.io/miro/
October 31, 2025 at 9:11 PM
Check out our new work: MIRO
No more post-training alignment!
We integrate human alignment right from the start, during pretraining!
Results:
✨ 19x faster convergence ⚡
✨ 370x less compute 💻
🔗 Explore the project: nicolas-dufour.github.io/miro/
No more post-training alignment!
We integrate human alignment right from the start, during pretraining!
Results:
✨ 19x faster convergence ⚡
✨ 370x less compute 💻
🔗 Explore the project: nicolas-dufour.github.io/miro/
Reposted by Lucas Degeorge
We introduce MIRO: a new paradigm for T2I model alignment integrating reward conditioning into pretraining, eliminating the need for separate fine-tuning/RL stages. This single-stage approach offers unprecedented efficiency and control.
- 19x faster convergence ⚡
- 370x less FLOPS than FLUX-dev 📉
- 19x faster convergence ⚡
- 370x less FLOPS than FLUX-dev 📉
October 31, 2025 at 11:24 AM
We introduce MIRO: a new paradigm for T2I model alignment integrating reward conditioning into pretraining, eliminating the need for separate fine-tuning/RL stages. This single-stage approach offers unprecedented efficiency and control.
- 19x faster convergence ⚡
- 370x less FLOPS than FLUX-dev 📉
- 19x faster convergence ⚡
- 370x less FLOPS than FLUX-dev 📉
Reposted by Lucas Degeorge
October 31, 2025 at 8:55 AM
Reposted by Lucas Degeorge
Very proud of our recent work, kudos to the team! Read @davidpicard.bsky.social’s excellent post for more details or the paper arxiv.org/pdf/2502.21318
October 8, 2025 at 9:19 PM
Very proud of our recent work, kudos to the team! Read @davidpicard.bsky.social’s excellent post for more details or the paper arxiv.org/pdf/2502.21318
Reposted by Lucas Degeorge
Final note: I'm (we're) tempted to organize a challenge on that topic as a workshop at a CV conf. ImageNet is the only source of images allowed and then you compete to get the bold numbers.
Do you think there would be people in for that? Do you think it would make for a nice competition?
Do you think there would be people in for that? Do you think it would make for a nice competition?
October 8, 2025 at 8:43 PM
Final note: I'm (we're) tempted to organize a challenge on that topic as a workshop at a CV conf. ImageNet is the only source of images allowed and then you compete to get the bold numbers.
Do you think there would be people in for that? Do you think it would make for a nice competition?
Do you think there would be people in for that? Do you think it would make for a nice competition?
Reposted by Lucas Degeorge
🚨Updated: "How far can we go with ImageNet for Text-to-Image generation?"
TL;DR: train a text2image model from scratch on ImageNet only and beat SDXL.
Paper, code, data available! Reproducible science FTW!
🧵👇
📜 arxiv.org/abs/2502.21318
💻 github.com/lucasdegeorg...
💽 huggingface.co/arijitghosh/...
TL;DR: train a text2image model from scratch on ImageNet only and beat SDXL.
Paper, code, data available! Reproducible science FTW!
🧵👇
📜 arxiv.org/abs/2502.21318
💻 github.com/lucasdegeorg...
💽 huggingface.co/arijitghosh/...
October 8, 2025 at 8:43 PM
🚨Updated: "How far can we go with ImageNet for Text-to-Image generation?"
TL;DR: train a text2image model from scratch on ImageNet only and beat SDXL.
Paper, code, data available! Reproducible science FTW!
🧵👇
📜 arxiv.org/abs/2502.21318
💻 github.com/lucasdegeorg...
💽 huggingface.co/arijitghosh/...
TL;DR: train a text2image model from scratch on ImageNet only and beat SDXL.
Paper, code, data available! Reproducible science FTW!
🧵👇
📜 arxiv.org/abs/2502.21318
💻 github.com/lucasdegeorg...
💽 huggingface.co/arijitghosh/...
Reposted by Lucas Degeorge
I had the privilege to be invited to speak about our work "Around the World in 80 Timesteps" at the French Podcast Underscore! If you speak french, i highly recommend it they did a great job with the montage!
If you want to learn more nicolas-dufour.github.io/plonk
www.youtube.com/watch?v=s5oH...
If you want to learn more nicolas-dufour.github.io/plonk
www.youtube.com/watch?v=s5oH...
Il a conçu la première IA d’OSINT (terrifiant… et génial)
YouTube video by Underscore_
www.youtube.com
July 31, 2025 at 4:43 PM
I had the privilege to be invited to speak about our work "Around the World in 80 Timesteps" at the French Podcast Underscore! If you speak french, i highly recommend it they did a great job with the montage!
If you want to learn more nicolas-dufour.github.io/plonk
www.youtube.com/watch?v=s5oH...
If you want to learn more nicolas-dufour.github.io/plonk
www.youtube.com/watch?v=s5oH...
Reposted by Lucas Degeorge
If you want to listen to Nicolas (in French) talking about generative models for geolocation, it's right now: m.twitch.tv/micode
Micode - Twitch
🌐 UNDERSCORE_
m.twitch.tv
June 25, 2025 at 6:25 PM
If you want to listen to Nicolas (in French) talking about generative models for geolocation, it's right now: m.twitch.tv/micode
🚨 News! 🚨
We have released the models from our latest paper "How far can we go with ImageNet for text-to-image generation?"
Check out the models on HuggingFace:
🤗 huggingface.co/Lucasdegeorg...
📜 arxiv.org/abs/2502.21318
We have released the models from our latest paper "How far can we go with ImageNet for text-to-image generation?"
Check out the models on HuggingFace:
🤗 huggingface.co/Lucasdegeorg...
📜 arxiv.org/abs/2502.21318
March 5, 2025 at 11:52 AM
🚨 News! 🚨
We have released the models from our latest paper "How far can we go with ImageNet for text-to-image generation?"
Check out the models on HuggingFace:
🤗 huggingface.co/Lucasdegeorg...
📜 arxiv.org/abs/2502.21318
We have released the models from our latest paper "How far can we go with ImageNet for text-to-image generation?"
Check out the models on HuggingFace:
🤗 huggingface.co/Lucasdegeorg...
📜 arxiv.org/abs/2502.21318
Reposted by Lucas Degeorge
Text-to-image models are trained on billions of data.
But, is it necessary?
Our "How far can we go with ImageNet for T2I generation?" @lucasdegeorge.bsky.social @arrijitghosh.bsky.social @nicolasdufour.bsky.social @davidpicard.bsky.social shows that no, if we are careful arxiv.org/abs/2502.21318
But, is it necessary?
Our "How far can we go with ImageNet for T2I generation?" @lucasdegeorge.bsky.social @arrijitghosh.bsky.social @nicolasdufour.bsky.social @davidpicard.bsky.social shows that no, if we are careful arxiv.org/abs/2502.21318
March 4, 2025 at 8:26 AM
Text-to-image models are trained on billions of data.
But, is it necessary?
Our "How far can we go with ImageNet for T2I generation?" @lucasdegeorge.bsky.social @arrijitghosh.bsky.social @nicolasdufour.bsky.social @davidpicard.bsky.social shows that no, if we are careful arxiv.org/abs/2502.21318
But, is it necessary?
Our "How far can we go with ImageNet for T2I generation?" @lucasdegeorge.bsky.social @arrijitghosh.bsky.social @nicolasdufour.bsky.social @davidpicard.bsky.social shows that no, if we are careful arxiv.org/abs/2502.21318
Reposted by Lucas Degeorge
Wow, neet! Reannotation is key here.
Conjecture:
As we are get more and more well-aligned text-image data, it will become easier and easier to train models.
This will allow us to explore both more streamlined and more exotic training recipes.
More signals that exciting times are coming!
Conjecture:
As we are get more and more well-aligned text-image data, it will become easier and easier to train models.
This will allow us to explore both more streamlined and more exotic training recipes.
More signals that exciting times are coming!
🚨 New preprint!
How far can we go with ImageNet for Text-to-Image generation? w. @arrijitghosh.bsky.social @lucasdegeorge.bsky.social @nicolasdufour.bsky.social @vickykalogeiton.bsky.social
TL;DR: Train a text-to-image model using 1000 less data in 200 GPU hrs!
📜https://arxiv.org/abs/2502.21318
🧵👇
How far can we go with ImageNet for Text-to-Image generation? w. @arrijitghosh.bsky.social @lucasdegeorge.bsky.social @nicolasdufour.bsky.social @vickykalogeiton.bsky.social
TL;DR: Train a text-to-image model using 1000 less data in 200 GPU hrs!
📜https://arxiv.org/abs/2502.21318
🧵👇
March 3, 2025 at 11:50 AM
Wow, neet! Reannotation is key here.
Conjecture:
As we are get more and more well-aligned text-image data, it will become easier and easier to train models.
This will allow us to explore both more streamlined and more exotic training recipes.
More signals that exciting times are coming!
Conjecture:
As we are get more and more well-aligned text-image data, it will become easier and easier to train models.
This will allow us to explore both more streamlined and more exotic training recipes.
More signals that exciting times are coming!
Reposted by Lucas Degeorge
These are some ridiculously good results from training tiny T2I models purely on ImageNet! It's almost too good to be true. Do check it out!
Check out our latest work on Text-to-Image generation! We've successfully trained a T2I model using only ImageNet data by leveraging captioning and data augmentation.
🚨 New preprint!
How far can we go with ImageNet for Text-to-Image generation? w. @arrijitghosh.bsky.social @lucasdegeorge.bsky.social @nicolasdufour.bsky.social @vickykalogeiton.bsky.social
TL;DR: Train a text-to-image model using 1000 less data in 200 GPU hrs!
📜https://arxiv.org/abs/2502.21318
🧵👇
How far can we go with ImageNet for Text-to-Image generation? w. @arrijitghosh.bsky.social @lucasdegeorge.bsky.social @nicolasdufour.bsky.social @vickykalogeiton.bsky.social
TL;DR: Train a text-to-image model using 1000 less data in 200 GPU hrs!
📜https://arxiv.org/abs/2502.21318
🧵👇
March 3, 2025 at 10:46 AM
These are some ridiculously good results from training tiny T2I models purely on ImageNet! It's almost too good to be true. Do check it out!
Reposted by Lucas Degeorge
🚨 New preprint!
How far can we go with ImageNet for Text-to-Image generation? w. @arrijitghosh.bsky.social @lucasdegeorge.bsky.social @nicolasdufour.bsky.social @vickykalogeiton.bsky.social
TL;DR: Train a text-to-image model using 1000 less data in 200 GPU hrs!
📜https://arxiv.org/abs/2502.21318
🧵👇
How far can we go with ImageNet for Text-to-Image generation? w. @arrijitghosh.bsky.social @lucasdegeorge.bsky.social @nicolasdufour.bsky.social @vickykalogeiton.bsky.social
TL;DR: Train a text-to-image model using 1000 less data in 200 GPU hrs!
📜https://arxiv.org/abs/2502.21318
🧵👇
March 3, 2025 at 10:19 AM
🚨 New preprint!
How far can we go with ImageNet for Text-to-Image generation? w. @arrijitghosh.bsky.social @lucasdegeorge.bsky.social @nicolasdufour.bsky.social @vickykalogeiton.bsky.social
TL;DR: Train a text-to-image model using 1000 less data in 200 GPU hrs!
📜https://arxiv.org/abs/2502.21318
🧵👇
How far can we go with ImageNet for Text-to-Image generation? w. @arrijitghosh.bsky.social @lucasdegeorge.bsky.social @nicolasdufour.bsky.social @vickykalogeiton.bsky.social
TL;DR: Train a text-to-image model using 1000 less data in 200 GPU hrs!
📜https://arxiv.org/abs/2502.21318
🧵👇
Reposted by Lucas Degeorge
🌍 Guessing where an image was taken is a hard, and often ambiguous problem. Introducing diffusion-based geolocation—we predict global locations by refining random guesses into trajectories across the Earth's surface!
🗺️ Paper, code, and demo: nicolas-dufour.github.io/plonk
🗺️ Paper, code, and demo: nicolas-dufour.github.io/plonk
December 10, 2024 at 3:56 PM
🌍 Guessing where an image was taken is a hard, and often ambiguous problem. Introducing diffusion-based geolocation—we predict global locations by refining random guesses into trajectories across the Earth's surface!
🗺️ Paper, code, and demo: nicolas-dufour.github.io/plonk
🗺️ Paper, code, and demo: nicolas-dufour.github.io/plonk