Shangyuan Tong, Haolin Jia, Hexiang Hu, Yu-Chuan Su, Mingda Zhang, XuanYang, Yandong Li, Xuhui Jia, and advisors Tommi Jaakkola and @saining.bsky.social! [n/n]
Shangyuan Tong, Haolin Jia, Hexiang Hu, Yu-Chuan Su, Mingda Zhang, XuanYang, Yandong Li, Xuhui Jia, and advisors Tommi Jaakkola and @saining.bsky.social! [n/n]
ArXiv: arxiv.org/pdf/2501.09732
Project Page: inference-scale-diffusion.github.io
ArXiv: arxiv.org/pdf/2501.09732
Project Page: inference-scale-diffusion.github.io
These results indicate that substantial training costs can be partially offset by modest inference-time compute, enabling higher-quality samples more efficiently. [6/n]
These results indicate that substantial training costs can be partially offset by modest inference-time compute, enabling higher-quality samples more efficiently. [6/n]
With the 12B FLUX.1-dev model on DrawBench, searching with all verifiers improves sample quality, while again specific improvement behaviors largely vary across different setups. [5/n]
With the 12B FLUX.1-dev model on DrawBench, searching with all verifiers improves sample quality, while again specific improvement behaviors largely vary across different setups. [5/n]
On ImageNet with SiT-XL, different combinations of verifiers and algorithms are observed to have very different scaling behaviors. [4/n]
On ImageNet with SiT-XL, different combinations of verifiers and algorithms are observed to have very different scaling behaviors. [4/n]
This suggests pushing the inference-time scaling limit by investing compute in searching for better noises.
Then, it's natural to ask: how do we know which sampling noises are good, and how do we search for such noises? [3/n]
This suggests pushing the inference-time scaling limit by investing compute in searching for better noises.
Then, it's natural to ask: how do we know which sampling noises are good, and how do we search for such noises? [3/n]
Therefore, for diffusion models to scale more at inference time, a new framework needs to be designed.[2/n]
Therefore, for diffusion models to scale more at inference time, a new framework needs to be designed.[2/n]