Existing dynamic scene interpolation methods typically assume that the motion between consecutive time steps is small enough so that displacements can be locally approximated by linear models. In practice, even slight deviations from this small-motion assumption can cause conventional techniques to fail. In this paper, we introduce Global Motion Corresponder (GMC), a novel approach that robustly handles large motion and achieves smooth transitions. GMC learns unary potential fields that predict SE(3) mappings into a shared canonical space, balancing correspondence, spatial and semantic smoothness, and local rigidity. We demonstrate that our method significantly outperforms existing baselines on 3D scene interpolation when the two states undergo large global motions. Furthermore, our method enables extrapolation capabilities where other baseline methods cannot.
For small inter-frame motion, local neighborhood searches yield correct correspondence and motion prediction
With large global motion, naïvely matching nearest neighbors results in criss-cross matches, which are unusable for scene interpolation.
An ideal method would be able to predict correct correspondence and achieve global motion.
We use Gaussian splats as our representations. First, two 3D Gaussian Splatting (3DGS) models are trained from the start state and end state. Then we transform both sets of Gaussians into a learnable shared canonical space where corresponding Gaussians occupy identical spatial locations:
$$ \underset{\hat{\boldsymbol{\mu}}^{(0)}_i}{\underbrace{\boldsymbol{R}^{(0)}_i \boldsymbol{\mu}_i^{(0)}+ \boldsymbol{t}^{(0)}_i}} = \underset{\hat{\boldsymbol{\mu}}^{(1)}_j}{\underbrace{\boldsymbol{R}^{(1)}_j \boldsymbol{\mu}_j^{(1)} + \boldsymbol{t}^{(1)}_{j}}}, \tag{1} $$where $\boldsymbol{R}$ and $\boldsymbol{t}$ represent learnable point-wise transformations for the two states. The transformations $(\boldsymbol{R}^{(0)}_i, \boldsymbol{t}^{(0)}_i)$ and $(\boldsymbol{R}^{(1)}_j, \boldsymbol{t}^{(1)}_j)$ are obtained from our Unary Potential Fields $\mathcal{F}_{0}$ and $\mathcal{F}_{1}$, which are parameterized as MLPs. The parameters of these MLPs are optimized using our proposed Energy-based loss. More details can be found in the paper.
Method Overview
Here are the scene interpolation results of our method, including both synthetic (row 1, 2) and real-world scenes (row 3, 4).
@misc{GlobalMotionCorresponder,
title={Global Motion Corresponder for 3D Point-Based Scene Interpolation under Large Motion},
author={Junru Lin and Chirag Vashist and Mikaela Angelina Uy and Colton Stearns and Xuan Luo and Leonidas Guibas and Ke Li},
year={2025},
eprint={2508.20136},
archivePrefix={arXiv},
primaryClass={eess.IV},
url={https://arxiv.org/abs/2508.20136},
}