Vladimir Yugay
banner
vyuga3d.bsky.social
Vladimir Yugay
@vyuga3d.bsky.social
Doing research in 3D Computer Vision.
Ph.D. student at the University of Amsterdam. Previously at TUM.

https://vladimiryugay.github.io/
Thanks to the team, Kien Nguyen, Theo Gevers, @cgmsnoek.bsky.social, and @martin-r-oswald.bsky.social from the University of Amsterdam!
October 7, 2025 at 9:02 AM
VoT does not require calibration or post-optimization and operates in real-time, capable of processing thousands of frames. It is trained on a vast amount of real-world indoor data, but can work just fine in outdoor scenarios. It uses only camera poses as supervision, making it broadly accessible
October 7, 2025 at 9:02 AM
📽️ Check out Visual Odometry Transformer! VoT is an end-to-end model for getting accurate metric camera poses from monocular videos.

vladimiryugay.github.io/vot/
October 7, 2025 at 9:02 AM
This work was conducted in collaboration wit Kersten Thies, @lucacarlone.bsky.social , Theo Gevers, @martin-r-oswald.bsky.social , and Lukas Schmid at the Computer Vision Group of the University of Amsterdam and the SPARKLab of @mit.edu
June 10, 2025 at 12:06 PM
We evaluate our method on synthetic and real-world datasets that undergo significant changes, including the movement, removal, and addition of large pieces of furniture, cutlery, a coffee machine, and pictures on the walls
June 10, 2025 at 12:06 PM
GaME detects scene changes and directly manipulates the 3D Gaussians to keep the map up to date. Additionally, our keyframe management system identifies and eliminates pixels that observe stale geometry, thereby minimizing the amount of discarded information
June 10, 2025 at 12:06 PM
We found two main problems. First, the 3D Gaussian maps can not easily “optimize out” changes in the geometry on the fly. Second, frames observing the old state of the scene contaminate the optimization process, resulting in visual artifacts and inconsistencies
June 10, 2025 at 12:06 PM
Imagine you want ot create a 3DGS map of your apartment. You reconstructed your kitchen and continued to the bedroom. While you are in the bedroom, someone has moved the chair and added a table in the kitchen without telling you. That’s what can happen with your reconstruction👇
June 10, 2025 at 12:06 PM
Introducing “Gaussian Mapping of Evolving Scenes”! We present an RGBD mapping system with novel view synthesis capabilities that accurately reconstruct scenes that change over time
vladimiryugay.github.io/game/
June 10, 2025 at 12:06 PM
Resubmission mentality in marathons

Munich 2023 -> 8 months prep -> COVID -> ❌

Amsterdam 2024 -> 6 months prep -> COVID -> ❌

Leiden 2025 -> 6 months prep -> lfg ✅
May 11, 2025 at 2:34 PM
⏩Code release for MAGiC-SLAM!
github.com/VladimirYuga...

We vibe-coded hard to make the code as simple as possible. Here are some features you can seamlessly integrate into your 3D reconstruction pipeline right away:
March 19, 2025 at 6:47 PM
This work was done with amazing collaborators Theo Gevers and @martin-r-oswald.bsky.social at the Computer Vision Group of the University of Amsterdam.
7/7
November 27, 2024 at 5:34 AM
Finally, we extend evaluation to novel view synthesis on real-world datasets. By extracting sequences from the ego-centric Aria dataset to simulate multi-agent operations, we prepared a hold-out test with novel view trajectories, ensuring a comprehensive evaluation of our system's capabilities.
6/7
November 27, 2024 at 5:34 AM
Our sub-maps inherently support local pose corrections provided by the loop closure module. Combined with an efficient caching scheme and a two-stage merging process, this allows for fast and precise global map reconstruction.
5/7
November 27, 2024 at 5:34 AM
Introducing “MAGiC-SLAM: Multi-Agent Gaussian Globally Consistent SLAM”! We do SLAM with novel view synthesis capabilities on multiple simultaneously operating agents!

vladimiryugay.github.io/magic_slam/i...
1/7
November 27, 2024 at 5:34 AM