SOTAVerified

MixSTE: Seq2seq Mixed Spatio-Temporal Encoder for 3D Human Pose Estimation in Video

2022-03-02CVPR 2022Code Available2· sign in to hype

Jinlu Zhang, Zhigang Tu, Jianyu Yang, Yujin Chen, Junsong Yuan

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Recent transformer-based solutions have been introduced to estimate 3D human pose from 2D keypoint sequence by considering body joints among all frames globally to learn spatio-temporal correlation. We observe that the motions of different joints differ significantly. However, the previous methods cannot efficiently model the solid inter-frame correspondence of each joint, leading to insufficient learning of spatial-temporal correlation. We propose MixSTE (Mixed Spatio-Temporal Encoder), which has a temporal transformer block to separately model the temporal motion of each joint and a spatial transformer block to learn inter-joint spatial correlation. These two blocks are utilized alternately to obtain better spatio-temporal feature encoding. In addition, the network output is extended from the central frame to entire frames of the input video, thereby improving the coherence between the input and output sequences. Extensive experiments are conducted on three benchmarks (Human3.6M, MPI-INF-3DHP, and HumanEva). The results show that our model outperforms the state-of-the-art approach by 10.9% P-MPJPE and 7.6% MPJPE. The code is available at https://github.com/JinluZhang1126/MixSTE.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Human3.6MMixSTE (HRNet, T=243)Average MPJPE (mm)39.8Unverified
Human3.6MMixSTE (CPN, T=243)Average MPJPE (mm)40.9Unverified
Human3.6MMixSTE (CPN, T=81)Average MPJPE (mm)42.4Unverified
HumanEva-IMixSTE (T=43, FT)Mean Reconstruction Error (mm)16.1Unverified
MPI-INF-3DHPMixSTE (T=27)MPJPE54.9Unverified
MPI-INF-3DHPMixSTE (T=1)MPJPE57.9Unverified

Reproductions