Ilyass Moummad
banner
ilyassmoummad.bsky.social
Ilyass Moummad
@ilyassmoummad.bsky.social
Postdoctoral Researcher @ Inria Montpellier (IROKO, Pl@ntNet)
SSL for plant images

Interested in Computer Vision, Natural Language Processing, Machine Listening, and Biodiversity Monitoring

Website: ilyassmoummad.github.io
[8/10] Semantically consistent retrieval 🔍
CroVCA retrieves correct classes even for fine-grained or ambiguous queries (e.g., indigo bird, grey langur).
✅ Outperforms Hashing-Baseline
✅ Works with only 16 bits and without supervision
November 3, 2025 at 2:31 PM
[7/10] Compact yet meaningful codes 💾
Even with just 16 bits, CroVCA preserves class structure.
t-SNE on CIFAR-10 shows clear, separable clusters — almost identical to the original 768-dim embeddings.
November 3, 2025 at 2:31 PM
[6/10] Strong performance across encoders 💪
Tested on multiple vision encoders (SimDINOv2, DINOv2, DFN…), CroVCA achieves SOTA unsupervised hashing:
November 3, 2025 at 2:30 PM
[4/10] HashCoder 🛠️
A lightweight MLP with final BatchNorm for balanced bits (inspired by OrthoHash). Can be used as:
🔹 Probe on frozen features
🔹 LoRA-based fine-tuning for efficient encoder adaptation
November 3, 2025 at 2:30 PM
[3/10] Unifying hashing 🔄
Can supervised + unsupervised hashing be done in one framework?

CroVCA aligns binary codes across semantically consistent views:
Augmentations → unsupervised
Class-consistent samples → supervised

🧩 One BCE loss + coding-rate regularizer
November 3, 2025 at 2:30 PM
They get rid off many tricks that help prevent collapse in DINO. Hoping that this will pave the way for more simplified solutions derived from principled approaches!
February 18, 2025 at 2:54 PM
By applying MaskCut on models trained with both DINO and SimDINO, they confirm similar emerging properties.
February 18, 2025 at 2:48 PM
They show that SimDINO is more stable during training as k-NN accuracy on ImageNet keeps improving. Additionally, when training on a different dataset (COCO), SimDINO is more robust.
February 18, 2025 at 2:42 PM
SimDINO and SimDINOv2 outperform DINO and DINOv2 both in classification and segmentation tasks. Additional experiments are provided in the paper.
February 18, 2025 at 2:33 PM
The regularization term measures the size of the covariance matrix of the embedding examples. It was proposed in 2007 by Ma to segment multivariate mixed data by minimizing the coding length (subject to distortion). Many great ideas and proofs can be found here: ieeexplore.ieee.org/document/428...
February 18, 2025 at 2:30 PM
Yi Ma & colleagues managed to simplify DINO & DINOv2 by removing many ingredients and adding a robust regularization term from information theory (coding rate) that learn informative decorrelated features. Happy to see principled approaches advance deep representation learning!
February 18, 2025 at 2:24 PM
A really cool paper from Kyutai demonstrates how model capabilities can be extended to a new domain (e.g., learning a new language) while preserving the original capabilities. This is achieved by leveraging the concept of adapters.
November 27, 2024 at 8:51 PM