We distill diffusion models into multiple 1-step students, allowing (a) improved quality by specializing in subsets and (b) improved latency by distilling into smaller architectures.
1/n
We distill diffusion models into multiple 1-step students, allowing (a) improved quality by specializing in subsets and (b) improved latency by distilling into smaller architectures.
1/n