🧩💻🗂️ All code, data, & checkpoints are released!
🔗 Learn more: jason-aplp.github.io/MOVIS/ (6/6)
🧩💻🗂️ All code, data, & checkpoints are released!
🔗 Learn more: jason-aplp.github.io/MOVIS/ (6/6)
🔹 Ours (with biased timestep scheduler) ✅
🔹 Zero123 (without it) ❌
Our approach shows more precise location prediction in the earlier stage & finer detail refinement in later stages! 🎯✨ (5/6)
🔹 Ours (with biased timestep scheduler) ✅
🔹 Zero123 (without it) ❌
Our approach shows more precise location prediction in the earlier stage & finer detail refinement in later stages! 🎯✨ (5/6)
📌 Larger timesteps → Focus on position & orientation recovery
📌 Smaller timesteps → Refine geometry & appearance
👇 We visualize the sampling process below! (3/6)
📌 Larger timesteps → Focus on position & orientation recovery
📌 Smaller timesteps → Refine geometry & appearance
👇 We visualize the sampling process below! (3/6)
🔍 Additional structural inputs (depth & mask)
🖌️ Novel-view mask prediction as an auxiliary task
🎯 A biased noise scheduler to facilitate training
We identify the following key insight: (2/6)
🔍 Additional structural inputs (depth & mask)
🖌️ Novel-view mask prediction as an auxiliary task
🎯 A biased noise scheduler to facilitate training
We identify the following key insight: (2/6)
Our model generalizes to in-the-wild scenes like YouTube videos🎥🌍! Using just *15 input views*, we achieve high-quality reconstructions with detailed geometry & appearance. 🌟 Watch the demo to see it in action! 👇 (5/n)
Our model generalizes to in-the-wild scenes like YouTube videos🎥🌍! Using just *15 input views*, we achieve high-quality reconstructions with detailed geometry & appearance. 🌟 Watch the demo to see it in action! 👇 (5/n)