Time of the Flight of the Gaussians: Optimizing Depth Indirectly in Dynamic Radiance Fields

Published: 11 Jun 2025, Last Modified: 13 Nov 2025CVPR 2025EveryoneCC BY 4.0
Abstract: We present a method to reconstruct dynamic scenes from monocular continuous-wave time-of-flight cameras using raw sensor samples that is as accurate as past methods and is 100$\times$ faster. Quickly achieving high-fidelity dynamic 3D reconstruction from a single viewpoint is a significant challenge in computer vision. Recent 3D Gaussian splatting methods often depend on multi-view data to produce satisfactory results and are brittle in their optimizations otherwise. In time-of-flight radiance field reconstruction, the property of interest---depth---is not directly optimized, causing additional challenges. We describe how these problems have a large and underappreciated impact upon the optimization when using a fast primitive-based scene representation like 3D Gaussians. Then, we incorporate two heuristics into our optimization to improve the accuracy of scene geometry for under-constrained time-of-flight Gaussians. Experimental results show that our approach produces accurate reconstructions under constrained sensing conditions, including for fast motions like swinging baseball bats.
Loading