Bálint Mucsányi
bmucsanyi.bsky.social
Bálint Mucsányi
@bmucsanyi.bsky.social
ELLIS & IMPRS-IS PhD Student at the University of Tübingen.

Excited about uncertainty quantification, weight spaces, and deep learning theory.
Yes, sampling is a possibility. For Dirichlets, the predictive, epistemic, and aleatoric estimators of the information-theoretical decomposition (Eq. (1) in the paper: arxiv.org/abs/2402.19460) are available in closed form, so we can do better than sampling (lines 3198 and 3234 of validate.py).
December 12, 2024 at 2:32 AM
The alphas are the parameters of the predictive Dirichlet distribution. If you normalize the alpha vector by the sum of its elements (called the "evidence"), you get the predictive mean which you can use for the ECE/MCE calculation. This is done automatically in line 3186 of validate.py. :)
December 12, 2024 at 2:24 AM
Correct! The EDLWrapper does not add extra parameters but changes the interpretation of logits, as the EDLLoss is a scheduled + regularized L2 loss instead of the usual cross-entropy.
December 12, 2024 at 2:20 AM
Of course, ask away! :)
December 11, 2024 at 1:51 AM
Many thanks to my amazing collaborators, @mkirchhof.bsky.social and @coallaoh.bsky.social!
December 3, 2024 at 1:38 PM
For more details, check out our paper: arxiv.org/abs/2402.19460! Our GitHub repo (github.com/bmucsanyi/un...) contains performant implementations of the 19 benchmarked uncertainty methods, out-of-the-box OOD perturbation support, handling of label uncertainty, and support for over 50 metrics. 7/7
December 3, 2024 at 1:38 PM
A promising avenue for disentanglement is to combine such specialized estimators. As a simple baseline, combining the Mahalanobis OOD detector's epistemic estimates and the aleatoric estimates of evidential methods leads to well-performing but only mildly correlated estimators. 6/7
December 3, 2024 at 1:38 PM
All these insights point to the conclusion that the best uncertainty method depends on the type of uncertainty and, even more importantly, the exact task we want to solve. Thinking only in terms of the 'aleatoric vs. epistemic' dichotomy is not fine-grained enough to obtain specialized methods. 5/7
December 3, 2024 at 1:38 PM
Predictive uncertainty encompasses all the aforementioned sources of uncertainty. Almost all methods perform well on predictive uncertainty metrics, but the best-performing one depends on the exact metric (see podiums below for different metrics). 4/7
December 3, 2024 at 1:38 PM
Instead, we found specialized estimators to perform best at capturing these sources of uncertainty. For epistemic uncertainty, a specialized OOD detector works best. For aleatoric uncertainty, evidential methods perform well, but more research is needed to develop dedicated aleatoric estimators. 3/7
December 3, 2024 at 1:38 PM
Decomposition formulas like in the image below are popular approaches for breaking up the total uncertainty into different parts. However, we unveil that these parts are severely internally correlated (rank corr. 0.8 to 0.999), i.e., they "measure the same thing" in practice. 2/7
December 3, 2024 at 1:38 PM
Could you add me to the starter pack? Thank you!
November 19, 2024 at 3:10 PM
Please add me. 😊
November 19, 2024 at 3:07 PM