SOTAVerified

CodingHomo: Bootstrapping Deep Homography With Video Coding

2025-04-16Code Available0· sign in to hype

Yike Liu, Haipeng Li, Shuaicheng Liu, Bing Zeng

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Homography estimation is a fundamental task in computer vision with applications in diverse fields. Recent advances in deep learning have improved homography estimation, particularly with unsupervised learning approaches, offering increased robustness and generalizability. However, accurately predicting homography, especially in complex motions, remains a challenge. In response, this work introduces a novel method leveraging video coding, particularly by harnessing inherent motion vectors (MVs) present in videos. We present CodingHomo, an unsupervised framework for homography estimation. Our framework features a Mask-Guided Fusion (MGF) module that identifies and utilizes beneficial features among the MVs, thereby enhancing the accuracy of homography prediction. Additionally, the Mask-Guided Homography Estimation (MGHE) module is presented for eliminating undesired features in the coarse-to-fine homography refinement process. CodingHomo outperforms existing state-of-the-art unsupervised methods, delivering good robustness and generalizability. The code and dataset are available at: githubhttps://github.com/liuyike422/CodingHomo

Tasks

Reproductions