ENS Paris. (prev ETH Zurich, Edinburgh, Oxford..)
Working on mathematical foundations/probabilistic interpretability of ML (what NNs learn🤷♂️, disentanglement🤔, king-man+woman=queen?👌…)
Full paper: arxiv.org/pdf/2410.22559
Full paper: arxiv.org/pdf/2410.22559
- max likelihood minimises
KL[p(x)||p’(x)] (p’(x)=model)
- max ELBO minimises
KL[p(x)q(z|x) || p’(x|z)p’(z)]
So brings together 2 models of the joint. (where p’(x)=\int p’(x|z)p’(z))
Can rearrange in diff ways, eg as
KL[p(x)q(z|x) || p’(x)p’(z|x)]
(or as in VAE)
- max likelihood minimises
KL[p(x)||p’(x)] (p’(x)=model)
- max ELBO minimises
KL[p(x)q(z|x) || p’(x|z)p’(z)]
So brings together 2 models of the joint. (where p’(x)=\int p’(x|z)p’(z))
Can rearrange in diff ways, eg as
KL[p(x)q(z|x) || p’(x)p’(z|x)]
(or as in VAE)