1st Place Solution of Multiview Egocentric Hand Tracking Challenge ECCV2024
2024-09-28Code Available2· sign in to hype
Minqiang Zou, Zhi Lv, Riqiang Jin, Tian Zhan, Mochen Yu, Yao Tang, Jiajun Liang
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/patiencefromzhou/simplehandpytorch★ 114
Abstract
Multi-view egocentric hand tracking is a challenging task and plays a critical role in VR interaction. In this report, we present a method that uses multi-view input images and camera extrinsic parameters to estimate both hand shape and pose. To reduce overfitting to the camera layout, we apply crop jittering and extrinsic parameter noise augmentation. Additionally, we propose an offline neural smoothing post-processing method to further improve the accuracy of hand position and pose. Our method achieves 13.92mm MPJPE on the Umetrack dataset and 21.66mm MPJPE on the HOT3D dataset.