Xi WANG
banner
xiwang92.bsky.social
Xi WANG
@xiwang92.bsky.social
Ecole Polytechnique, IP Paris; Prev. Ph.D.@Univ Rennes, Inria/IRISA
https://triocrossing.github.io/
Our approach fundamentally differs from previous distillation methods, such as DMD. Instead of minimizing the divergence of denoising distributions across the entire latent space, Di[M]O optimizes the divergence of token-level conditional distributions.
March 21, 2025 at 3:36 PM
To approximate the loss gradient, we introduce an auxiliary model that estimates an otherwise intractable term in the loss function. The auxiliary model is trained using a standard MDM training loss, with one-step generated samples as targets.
March 21, 2025 at 3:36 PM
To sample from the correct joint distribution, we introduce an initialization that maps a randomized input sequence to an almost deterministic target sequence.
Without proper initialization, the model may suffer from divergence or mode collapse, making this step essential.
March 21, 2025 at 3:36 PM
The key idea is inspired by on-policy distillation. We align the output distributions of the teacher and student models at the student generated intermediate states, ensuring that the student's generation closely matches the teacher's by covering all possible intermediate states.
March 21, 2025 at 3:36 PM
Masked Diffusion Models (MDMs) are a hot topic in generative AI 🔥 — powerful but slow due to multiple sampling steps.
We @polytechniqueparis.bsky.social and @inria-grenoble.bsky.social introduce Di[M]O — a novel approach to distill MDMs into a one-step generator without sacrificing quality.
March 21, 2025 at 3:36 PM