These results indicate that substantial training costs can be partially offset by modest inference-time compute, enabling higher-quality samples more efficiently. [6/n]
These results indicate that substantial training costs can be partially offset by modest inference-time compute, enabling higher-quality samples more efficiently. [6/n]
With the 12B FLUX.1-dev model on DrawBench, searching with all verifiers improves sample quality, while again specific improvement behaviors largely vary across different setups. [5/n]
With the 12B FLUX.1-dev model on DrawBench, searching with all verifiers improves sample quality, while again specific improvement behaviors largely vary across different setups. [5/n]
On ImageNet with SiT-XL, different combinations of verifiers and algorithms are observed to have very different scaling behaviors. [4/n]
On ImageNet with SiT-XL, different combinations of verifiers and algorithms are observed to have very different scaling behaviors. [4/n]
This suggests pushing the inference-time scaling limit by investing compute in searching for better noises.
Then, it's natural to ask: how do we know which sampling noises are good, and how do we search for such noises? [3/n]
This suggests pushing the inference-time scaling limit by investing compute in searching for better noises.
Then, it's natural to ask: how do we know which sampling noises are good, and how do we search for such noises? [3/n]
In our latest study—Inference-Time Scaling for Diffusion Models beyond Scaling Denoising Steps—we reframe inference-time scaling as a search problem over sampling noises. 🧵[1/n]
In our latest study—Inference-Time Scaling for Diffusion Models beyond Scaling Denoising Steps—we reframe inference-time scaling as a search problem over sampling noises. 🧵[1/n]