- We propose a mixed residual-based representation for dense RGB-D reconstruction of large-scale scenes, which preserves fine-grained details with relatively low memory and computational cost.
- We propose a residual-based bundle adjustment technique that employs a tiny MLP for residual-based pose refinement. Compared to traditional BAs, our method improves pose estimation in terms of both efficiency and robustness.
- We have implemented an efficient system of online RGB-D dense reconstruction which realizes robust and fine-grained real-time reconstruction for large scenes over 1000𝑚2 with an affordable GPU memory footprint.
Representative Image. We present RemixFusion, a residual-based RGB-D framework by virtue of both explicit and implicit representations for large-scale online dense
reconstruction. RemixFusion can support real-time fine-grained reconstruction in a memory-efficient way. It only costs 9.8GB GPU memory with 12 FPS for
the about 400𝑚2 reconstruction above, while other methods [Johari et al. 2023; Tang et al. 2023; Zhu et al. 2022] struggle in both tracking and reconstruction in
real time. Traditional explicit methods fail for this scene. GS-ICP SLAM [Ha et al. 2024] is the SOTA 3DGS-based SLAM. The average results of reconstruction
and tracking on the BS3D dataset as well as the system FPS and GPU memory usage on the above scene are shown on the right, which illustrates RemixFusion
obtains better performance and efficiency. RemixFusion-lite denotes the lightweight version and achieves decent performance with about 25 FPS.