Jianyuan Wang
@jianyuanwang.bsky.social
I am trying to. Probably we could hear about this around next submission ddl 😂
March 18, 2025 at 4:51 AM
I am trying to. Probably we could hear about this around next submission ddl 😂
It seems so (with a short glance only). The techniques used by Fast3R can also be applied to VGGT
March 18, 2025 at 4:49 AM
It seems so (with a short glance only). The techniques used by Fast3R can also be applied to VGGT
Haha, this probably serves as an indirect validation of NVIDIA’s stock value.
March 18, 2025 at 4:48 AM
Haha, this probably serves as an indirect validation of NVIDIA’s stock value.
Currently, this training approach is not very stable, but I believe that’s likely because I haven’t yet found the correct training method. I hope this can achieve better results in the future, which could then avoid an explicit modelling of point map.
March 17, 2025 at 12:40 PM
Currently, this training approach is not very stable, but I believe that’s likely because I haven’t yet found the correct training method. I hope this can achieve better results in the future, which could then avoid an explicit modelling of point map.
Finally, great work together with Minghao Chen, Nikita Karaev, Andrea Vedaldi, Christian Rupprecht, and David Novotny!
@oxford-vgg.bsky.social
@oxford-vgg.bsky.social
March 17, 2025 at 2:14 AM
Finally, great work together with Minghao Chen, Nikita Karaev, Andrea Vedaldi, Christian Rupprecht, and David Novotny!
@oxford-vgg.bsky.social
@oxford-vgg.bsky.social
Interesting observation: VGGT’s camera & depth predictions are highly accurate and consistent. Unprojecting our predicted depth with predicted camera parameters yields even more precise point clouds than directly predicted point maps! Try this yourself using the Hugging Face demo 🤗
March 17, 2025 at 2:12 AM
Interesting observation: VGGT’s camera & depth predictions are highly accurate and consistent. Unprojecting our predicted depth with predicted camera parameters yields even more precise point clouds than directly predicted point maps! Try this yourself using the Hugging Face demo 🤗
Compared to concurrent CVPR'25 Transformer-based 3D reconstruction methods, VGGT achieves significantly higher accuracy, with speed similar to the fastest variant Fast3R.
March 17, 2025 at 2:12 AM
Compared to concurrent CVPR'25 Transformer-based 3D reconstruction methods, VGGT achieves significantly higher accuracy, with speed similar to the fastest variant Fast3R.
Bonus insight: Using pretrained VGGT significantly enhances downstream tasks like:
🚀 Non-rigid point tracking
🚀 Feed-forward novel view synthesis
🚀 Non-rigid point tracking
🚀 Feed-forward novel view synthesis
March 17, 2025 at 2:12 AM
Bonus insight: Using pretrained VGGT significantly enhances downstream tasks like:
🚀 Non-rigid point tracking
🚀 Feed-forward novel view synthesis
🚀 Non-rigid point tracking
🚀 Feed-forward novel view synthesis
A strong advantage of our method is the ability to predict 3D attributes without any expensive optimization. For example, 🔸 VGGT can easily process ~200 images in ~10s on a single 40GB A100 GPU 🔸 50x faster than optimization-based methods, using far less memory.
March 17, 2025 at 2:11 AM
A strong advantage of our method is the ability to predict 3D attributes without any expensive optimization. For example, 🔸 VGGT can easily process ~200 images in ~10s on a single 40GB A100 GPU 🔸 50x faster than optimization-based methods, using far less memory.
Try our demo live on Hugging Face Spaces!
🤗: huggingface.co/spaces/faceb...
(See demo illustration below) 👇
🤗: huggingface.co/spaces/faceb...
(See demo illustration below) 👇
March 17, 2025 at 2:10 AM
Try our demo live on Hugging Face Spaces!
🤗: huggingface.co/spaces/faceb...
(See demo illustration below) 👇
🤗: huggingface.co/spaces/faceb...
(See demo illustration below) 👇
No expensive optimization needed, yet delivers SOTA results for:
✅ Camera Pose Estimation
✅ Multi-view Depth Estimation
✅ Dense Point Cloud Reconstruction
✅ Point Tracking
✅ Camera Pose Estimation
✅ Multi-view Depth Estimation
✅ Dense Point Cloud Reconstruction
✅ Point Tracking
March 17, 2025 at 2:08 AM
No expensive optimization needed, yet delivers SOTA results for:
✅ Camera Pose Estimation
✅ Multi-view Depth Estimation
✅ Dense Point Cloud Reconstruction
✅ Point Tracking
✅ Camera Pose Estimation
✅ Multi-view Depth Estimation
✅ Dense Point Cloud Reconstruction
✅ Point Tracking