We present Reusable Motion prior (ReMP), an effective motion prior that can accurately track the temporal evolution of motion in various downstream tasks. Inspired by the success of foundation models, we argue that a robust spatio-temporal motion prior can encapsulate underlying 3D dynamics applicable to various sensor modalities. We learn the rich motion prior from a sequence of complete parametric models of posed human body shape. Our prior can easily estimate poses in missing frames or noisy measurements despite significant occlusion by employing a temporal attention mechanism. More interestingly, our prior can guide the system with incomplete and challenging input measurements to quickly extract critical information to estimate the sequence of poses, significantly improving the training efficiency for mesh sequence recovery. ReMP consistently outperforms the baseline method on diverse and practical 3D motion data, including depth point clouds, LiDAR scans, and IMU sensor data.
The overall pipeline of our method consists of two parts: (a) training motion prior and (b) reusing pre-trained prior. In the motion prior training phase, a sequence of pose parameters θ and the root translation transitions Δx form a sequence of motion parameter M. We use a transformer encoder and MLP layers to generate Gaussian distributions where we can sample the latent vectors. We feed the latent vectors to a transformer decoder to generate the motion parameters then to the SMPL parameters. After training the prior, we freeze all the networks used in the first phase. In the reusing phase, we encode the input data and use a transformer encoder to generate a distribution that is then used to sample the latent vectors for the transformer decoder. We use an additional shape parameter estimator for β. Finally, we combine all three parameters with the SMPL layer to reconstruct the human motion.
Input | GT | ReMP | DMR[1] | VoteHMR[2] | Zuo et al.[3] |
---|
Input | ReMP | DMR[1] | VoteHMR[2] | Zuo et al.[3] |
---|
Input | GT | ReMP | DMR[1] | VoteHMR[2] | Zuo et al.[3] |
---|
GT | ReMP | PIP[4] | TransPose[5] |
---|
@InProceedings{Jang_2025_WACV, author = {Jang, Hojun and Kim, Young Min}, title = {ReMP: Reusable Motion Prior for Multi-domain 3D Human Pose Estimation and Motion Inbetweening}, booktitle = {Proceedings of IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {}, year = {2025}, pages = {} }