W: https://zubairirshad.com
Please check these works out if you haven’t already!
Please check these works out if you haven’t already!
This was a fun collaboration with
@vitorguizilini, @SashaKhazatsky and @KarlPertsch!
This was a fun collaboration with
@vitorguizilini, @SashaKhazatsky and @KarlPertsch!
• Extending to in-the-wild scenes via foundation models for robot segmentation & keypoints.
• Ensembling predictions over time for better temporal consistency.
• Fine-tuning pointmap models on real robot data to handle cluttered tabletops.
8/n
• Extending to in-the-wild scenes via foundation models for robot segmentation & keypoints.
• Ensembling predictions over time for better temporal consistency.
• Fine-tuning pointmap models on real robot data to handle cluttered tabletops.
8/n
• CtRNet-X is trained on Panda; generalization to other robots is untested.
• DUSt3R struggles with clutter or minimal view overlap.
• Steps 2️⃣ & 3️⃣ may yield false positives in tough lighting or geometry.
7/n
• CtRNet-X is trained on Panda; generalization to other robots is untested.
• DUSt3R struggles with clutter or minimal view overlap.
• Steps 2️⃣ & 3️⃣ may yield false positives in tough lighting or geometry.
7/n
6/n
6/n
1️⃣ and 2️⃣ quality metrics show IOU and Reprojection-error distributions post-calibration.
5/n
1️⃣ and 2️⃣ quality metrics show IOU and Reprojection-error distributions post-calibration.
5/n
4/n
4/n
🤖 ~36k calibrated episodes with good-quality extrinsic calibration
🦾 ~24k calibrated multi-view episodes with good-quality multi-view camera calibration
✅ Quality assessment metrics for all provided camera poses
3/n
🤖 ~36k calibrated episodes with good-quality extrinsic calibration
🦾 ~24k calibrated multi-view episodes with good-quality multi-view camera calibration
✅ Quality assessment metrics for all provided camera poses
3/n
1️⃣ Auto Segment Anything (SAM) based filtering (Camera-to-Base Calibration)
2️⃣ Tuned CtRNet-X for bringing in additional cams (Camera-to-Base Calibration)
3️⃣ Pretrained DUST3R with depth-based pose optimization (Camera-to-Camera Calibration)
2/n
1️⃣ Auto Segment Anything (SAM) based filtering (Camera-to-Base Calibration)
2️⃣ Tuned CtRNet-X for bringing in additional cams (Camera-to-Base Calibration)
3️⃣ Pretrained DUST3R with depth-based pose optimization (Camera-to-Camera Calibration)
2/n
Join us in shaping the future of robotics, 3D vision, and language models! 🤖📚 #CVPR2025
Join us in shaping the future of robotics, 3D vision, and language models! 🤖📚 #CVPR2025
⭐ Angel Chang (Simon Fraser University)
⭐ Chelsea Finn (Stanford University)
⭐ Hao Su (UC San Diego)
⭐ Katerina Fragkiadaki (CMU)
⭐ Yunzhu Li (Columbia University)
⭐ Ranjay Krishna (University of Washington)
5/N
⭐ Angel Chang (Simon Fraser University)
⭐ Chelsea Finn (Stanford University)
⭐ Hao Su (UC San Diego)
⭐ Katerina Fragkiadaki (CMU)
⭐ Yunzhu Li (Columbia University)
⭐ Ranjay Krishna (University of Washington)
5/N
✅ 3D Vision-Language Policy Learning
✅ Pretraining for 3D VLMs
✅ 3D Representations for Policy Learning
✅ 3D Benchmarks & Simulation Frameworks
✅ 3D Vision-Language Action Models
✅ 3D Instruction-Tuning & Pretraining Datasets for Robotics
4/N
✅ 3D Vision-Language Policy Learning
✅ Pretraining for 3D VLMs
✅ 3D Representations for Policy Learning
✅ 3D Benchmarks & Simulation Frameworks
✅ 3D Vision-Language Action Models
✅ 3D Instruction-Tuning & Pretraining Datasets for Robotics
4/N
📅 Deadline: April 15, 2024 (11:59 PM PST)
📜 Format: Up to 4 pages (excluding references/appendices), CVPR template, anonymized submissions
🏆 Accepted papers: Poster presentations, with selected papers receiving spotlight talks!
3/N
📅 Deadline: April 15, 2024 (11:59 PM PST)
📜 Format: Up to 4 pages (excluding references/appendices), CVPR template, anonymized submissions
🏆 Accepted papers: Poster presentations, with selected papers receiving spotlight talks!
3/N
2/N
2/N