Street Gaussians without 3D Object Tracker ICCV 2025
- Ruida Zhang1,2
- Chengxi Li1
- Chenyangguang Zhang1
- Xingyu Liu1
- Haili Yuan1
- Yanyan Li2
- Xiangyang Ji1
- Gim Hee Lee2
- Tsinghua University1
- National University of Singapore2

Comparison of 3D tracker-based Street Gaussians (left) and our approach (right). Existing methods heavily rely on object poses, but 3D trackers struggle with limited generalization, leading to flaws in novel view synthesis. In contrast, 2D foundation models show better generalization. Our approach leverages a 2D foundation model for object tracking and learns point motion within an implicit feature space to autonomously correct tracking errors, improving robustness across diverse scenes.
Abstract
Realistic scene reconstruction in driving scenarios poses significant challenges due to fast-moving objects. Most existing methods rely on labor-intensive manual labeling of object poses to reconstruct dynamic objects in canonical space and move them based on these poses during rendering. While some approaches attempt to use 3D object trackers to replace manual annotations, the limited generalization of 3D trackers - caused by the scarcity of large-scale 3D datasets - results in inferior reconstructions in real-world settings. In contrast, 2D foundation models demonstrate strong generalization capabilities. To eliminate the reliance on 3D trackers and enhance robustness across diverse environments, we propose a stable object tracking module by leveraging associations from 2D deep trackers within a 3D object fusion strategy. We address inevitable tracking errors by further introducing a motion learning strategy in an implicit feature space that autonomously corrects trajectory errors and recovers missed detections. Experimental results on Waymo-NOTR and KITTI show that our method outperforms existing approaches.
Method Overview

Comparisons
