Luigi Acerbi
banner
lacerbi.bsky.social
Luigi Acerbi
@lacerbi.bsky.social
Assoc. Prof. of Machine & Human Intelligence | Univ. Helsinki & Finnish Centre for AI (FCAI) | Bayesian ML & probabilistic modeling | https://lacerbi.github.io/
I was thinking of sharing them publicly on my blog but then realized it could be problematic / misunderstood / misused, so for now I refrained from it... but yes I can send them privately. I'll email you later.
November 12, 2025 at 9:40 AM
We regularly use LLMs to review our own papers before submission. I agree that in principle it's hard to "prove" that a review is entirely LLM-generated vs. LLM-written... but in practice it's quite easy especially when you see multiple "LLM-smelling" points in a review.
November 12, 2025 at 9:26 AM
AISTATS adopted the iron fist against "LLM-generated" reviews and I fully agree with it.

Hopefully we can figure out a way to allow responsible, "LLM-assisted" reviews where one genuinely uses the LLM to assist (not replace) the reviewer, which can be a force multiplier.
November 12, 2025 at 9:16 AM
Even just the fact that we have oral tests (throughout school and then university) shocks a lot of people.

It's easy to fall into the fallacy "I went through this system so it is obviously the best system" and surely it has downsides but tbh it does prepare you for a lot of things in life...
November 3, 2025 at 10:34 AM
I fully agree. I have been thinking a lot about this and my current advice is use the standard error of the median (bootstrapped), with a separate metric for reliability if needed. The latter is less standard, we have been reporting e.g. 80% performance quantile, and std err of that (bootstrapped)
October 26, 2025 at 12:47 PM
All correct. A steelman argument can be made that if you are comparing methods you also may care about the actual dispersion which is somewhat related to reliability. But even then you‘d probably best showcase other metrics focusing on that. And yes ML would benefit from more stats.
October 26, 2025 at 11:52 AM
THATS TEH SPIRIT!!!
October 1, 2025 at 8:21 PM
Kudos to fellow co-organizers! @desirivanova.bsky.social
@conorhassan.bsky.social Cen-You (Scott) Li
October 1, 2025 at 1:57 PM
Short abstract submission is open (DL Oct 16, 2025 AoE)
Accepted abstracts will be presented as posters.

🌍 Location: Copenhagen (co-located with #EurIPS)
📅 Date: 2 Dec, 2025
🔗 Website: sites.google.com/view/amortiz...
📝Call for abstracts: sites.google.com/view/amortiz...
Amortized ProbML 2025 - Call for Papers
Accepted papers will be presented as posters during the workshop poster session on 2nd December, 2025. Please be aware that this is not the same as the ELLIS UnConference poster session.
sites.google.com
October 1, 2025 at 1:57 PM
We'll discuss tricky questions such as:
• When does amortization fail badly even though it's rarely talked about?
• Which applications can these neat methods actually work for?
• Can we really deploy these methods in the real world?
• ... and more!
October 1, 2025 at 1:57 PM
Very interesting, looking forward to reading it in detail! For broader context, I recommend looking into the woefully underappreciated literature on (autoregressive) (transformer) neural processes. (Conditional) NPs are probabilistic models predicting data based on arbitrary sets of data.
September 16, 2025 at 4:37 AM
For example for surrogate-based Bayesian inference, our VBMC works with GPs in the low-likelihood evaluation regime:
github.com/acerbilab/py...

Then for many likelihood evaluations we have other methods, like this one with normalizing flows:
arxiv.org/abs/2504.11554
Normalizing Flow Regression for Bayesian Inference with Offline Likelihood Evaluations
Bayesian inference with computationally expensive likelihood evaluations remains a significant challenge in many scientific domains. We propose normalizing flow regression (NFR), a novel offline infer...
arxiv.org
July 20, 2025 at 12:11 PM
Yep!

We are not quite doing it yet in a single existing method -- mostly because it'd be hard to publish --, but definitely thought about it many times.
July 20, 2025 at 12:11 PM
I assume it was referring to different authors of multi-author papers. I also read it as boilerplate “we are not competing with NeurIPS, also go to NeurIPS” by making it clear that this is just a satellite and not a full-blown schism (plus there likely isn’t enough space for everyone).
July 17, 2025 at 5:10 AM