And if it's not, it is differentiable everywhere.
And if it's not, we avoid the kinks almost surely.
And if we don't, what is computed is a subgradient.
And if it's not, it approximates one.
And if that's not true, who cares? The loss went down.
Adaptive Conditional Gradient Descent
https://arxiv.org/abs/2510.11440
Adaptive Conditional Gradient Descent
https://arxiv.org/abs/2510.11440
**Differentiable Generalized Sliced Wasserstein Plans**
w/
L. Chapel
@rtavenar.bsky.social
We propose a Generalized Sliced Wasserstein method that provides an approximated transport plan and which admits a differentiable approximation.
arxiv.org/abs/2505.22049 1/5
functions" is accepted at Mathematical Programming !! This is a joint work with Jerome Bolte, Eric Moulines and Edouard Pauwels where we study a subgradient method with errors for nonconvex nonsmooth functions.
arxiv.org/pdf/2404.19517
functions" is accepted at Mathematical Programming !! This is a joint work with Jerome Bolte, Eric Moulines and Edouard Pauwels where we study a subgradient method with errors for nonconvex nonsmooth functions.
arxiv.org/pdf/2404.19517
And if it's not, it is differentiable everywhere.
And if it's not, we avoid the kinks almost surely.
And if we don't, what is computed is a subgradient.
And if it's not, it approximates one.
And if that's not true, who cares? The loss went down.
And if it's not, it is differentiable everywhere.
And if it's not, we avoid the kinks almost surely.
And if we don't, what is computed is a subgradient.
And if it's not, it approximates one.
And if that's not true, who cares? The loss went down.
Check out SCION: a new optimizer that adapts to the geometry of your problem using norm-constrained linear minimization oracles (LMOs): 🧵👇
We got this idea after their cool work on improving Plug and Play with FM: arxiv.org/abs/2410.02423
We got this idea after their cool work on improving Plug and Play with FM: arxiv.org/abs/2410.02423