mrnerf.bsky.social
@mrnerf.bsky.social
December 6, 2024 at 7:58 AM
- Our HybridGS achieves state-of-the-art performance in benchmark datasets, outperforming previous methods and setting a new standard for novel view synthesis in scenes with transients.
December 6, 2024 at 7:58 AM
- We develop a multi-view supervision scheme for 3DGS that utilizes overlapping regions across multiple views. This enhances the model’s capability to distinguish between static and transient elements, ultimately improving the overall quality of the novel view synthesis.
December 6, 2024 at 7:58 AM
**Contributions:**
- We are the first to introduce a novel hybrid representation that combines image-specific 2D Gaussians with static 3D Gaussians, enabling effective modeling of transient objects within casually captured images.
December 6, 2024 at 7:58 AM
December 6, 2024 at 7:43 AM
• On various challenging real-world dynamic scenes, we surpass existing state-of-the-art approaches on all metrics: reconstruction quality, memory utilization, and training and rendering speed.
December 6, 2024 at 7:43 AM
• We introduce a learned quantization-sparsity framework for compressing per-frame residuals, initializing and training it efficiently using viewspace gradient differences that separate dynamic and static scene content.
December 6, 2024 at 7:43 AM
Contributions:
• We propose a Gaussian residual-based framework to model 3D dynamic scenes for online FVV without any structural constraints. This allows free learning of all 3D-GS attribute residuals, resulting in higher model expressiveness.
December 6, 2024 at 7:43 AM
Additionally, our neural-free sparse voxels are seamlessly compatible with grid-based 3D processing algorithms.

We achieve promising mesh reconstruction accuracy by integrating TSDF-Fusion and Marching Cubes into our sparse grid system.
December 6, 2024 at 7:34 AM
Our method improves the previous neural-free voxel grid representation by over 4db PSNR and more than 10x rendering FPS speedup, achieving state-of-the-art comparable novel-view synthesis results.
December 6, 2024 at 7:34 AM
This avoids the well-known popping artifact found in Gaussian splatting. Second, we adaptively fit sparse voxels to different levels of detail within scenes, faithfully reproducing scene details while achieving high rendering frame rates.
December 6, 2024 at 7:34 AM
There are two key contributions coupled with the proposed system. The first is to render sparse voxels in the correct depth order along pixel rays by using dynamic Morton ordering.
December 6, 2024 at 7:34 AM
December 6, 2024 at 7:29 AM
Empirically, we find that their rank order is well-defined in synthetic data, but the complexity of real-world data currently overwhelms the differences. Furthermore, the fast rendering speed of all Gaussian-based methods comes at the cost of brittleness in optimization.
December 6, 2024 at 7:29 AM
We use multiple existing datasets and a new instructive synthetic dataset designed to isolate factors that affect reconstruction quality. We systematically categorize Gaussian splatting methods into specific motion representation types and quantify how their differences impact performance.
December 6, 2024 at 7:29 AM
In this work, we organize, benchmark, and analyze many Gaussian-splatting-based methods, providing apples-to-apples comparisons that prior works have lacked.
December 6, 2024 at 7:29 AM