Member of @belongielab.org, ELLIS @ellis.eu, and Pioneer Centre for AI🤖
Computer Vision | Multimodality
MSc CS at ETH Zurich
🔗: zhaochongan.github.io/
🔗 Code: github.com/ZhaochongAn/...
#ICLR2025 #Multimodality #3DSegmentation @belongielab.org @ellis.eu @ethzurich.bsky.social @ox.ac.uk
🔗 Code: github.com/ZhaochongAn/...
#ICLR2025 #Multimodality #3DSegmentation @belongielab.org @ellis.eu @ethzurich.bsky.social @ox.ac.uk
📝: arxiv.org/pdf/2410.22489
💻: github.com/ZhaochongAn/Multimodality-3D-Few-Shot
📝: arxiv.org/pdf/2410.22489
💻: github.com/ZhaochongAn/Multimodality-3D-Few-Shot
💠 2D images (leveraged implicitly during pretraining)
💠 Text (using class names)
—all at no extra cost beyond the 3D-only setup. ✨
💠 2D images (leveraged implicitly during pretraining)
💠 Text (using class names)
—all at no extra cost beyond the 3D-only setup. ✨
However, when support and query objects look very different, performance can suffer, limiting effective few-shot adaptation. 🙁
However, when support and query objects look very different, performance can suffer, limiting effective few-shot adaptation. 🙁
With MM-FSS, we take it even further!
Ref: COSeg arxiv.org/pdf/2410.22489
With MM-FSS, we take it even further!
Ref: COSeg arxiv.org/pdf/2410.22489