Fredrik K. Gustafsson
fregu856.bsky.social
Fredrik K. Gustafsson
@fregu856.bsky.social
Postdoc at IBME in Oxford. Machine learning for healthcare.
https://www.fregu856.com/
Isn't there a lot of noise in these decisions, just like for conference papers etc?
October 31, 2025 at 4:04 PM
Grattis!
October 10, 2025 at 4:31 AM
The waiting area is also quite dull and gets really crowded, probably the worst part of my entire trip.
July 16, 2025 at 10:08 AM
I think it was already in the batch of papers I was given to rate, basically no pathology-related papers, for example. ICML was definitely better in this regard.
July 4, 2025 at 4:16 AM
Looks very useful, thanks for sharing!
April 2, 2025 at 6:57 AM
Nice, saw this on arxiv and thought it seemed interesting, might read this as well, thanks!
March 24, 2025 at 7:42 AM
Nice, thanks!

I actually wrote "The one proper method change that seems to have the biggest effect is probably adding the KoLeo regularization loss term?" in my notes, so would be nice to read more about how that works.
March 23, 2025 at 3:46 PM
The main thing definitely seems to be that they scale iBOT from ViT-L/16 trained on ImageNet-22k (14 million images) to ViT-g/14 trained on their LVD-142M dataset (142 million images).

Their model distillation approach is also interesting, distilling their ViT-g down to ViT-L and smaller models.
March 23, 2025 at 3:21 PM
"We revisit existing discriminative self-supervised approaches [...] such as iBOT, and we reconsider some of their design choices under the lens of a larger dataset. Most of our technical contributions are tailored toward stabilizing and accelerating [...] when scaling in model and data sizes"
March 23, 2025 at 3:21 PM
iBOT: Image BERT Pre-Training with Online Tokenizer (ICLR 2022)

DINOv2: Learning Robust Visual Features without Supervision (TMLR, 2024)

DINOv2 doesn't really add much methodological difference compared to iBOT, they give a good summary of what they do:
March 23, 2025 at 3:21 PM