SOTAVerified

Focus on Local: Finding Reliable Discriminative Regions for Visual Place Recognition

2025-04-14Code Available1· sign in to hype

Changwei Wang, Shunpeng Chen, Yukun Song, Rongtao Xu, Zherui Zhang, Jiguang Zhang, Haoran Yang, Yu Zhang, Kexue Fu, Shide Du, Zhiwei Xu, Longxiang Gao, Li Guo, Shibiao Xu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Visual Place Recognition (VPR) is aimed at predicting the location of a query image by referencing a database of geotagged images. For VPR task, often fewer discriminative local regions in an image produce important effects while mundane background regions do not contribute or even cause perceptual aliasing because of easy overlap. However, existing methods lack precisely modeling and full exploitation of these discriminative regions. In this paper, we propose the Focus on Local (FoL) approach to stimulate the performance of image retrieval and re-ranking in VPR simultaneously by mining and exploiting reliable discriminative local regions in images and introducing pseudo-correlation supervision. First, we design two losses, Extraction-Aggregation Spatial Alignment Loss (SAL) and Foreground-Background Contrast Enhancement Loss (CEL), to explicitly model reliable discriminative local regions and use them to guide the generation of global representations and efficient re-ranking. Second, we introduce a weakly-supervised local feature training strategy based on pseudo-correspondences obtained from aggregating global features to alleviate the lack of local correspondences ground truth for the VPR task. Third, we suggest an efficient re-ranking pipeline that is efficiently and precisely based on discriminative region guidance. Finally, experimental results show that our FoL achieves the state-of-the-art on multiple VPR benchmarks in both image retrieval and re-ranking stages and also significantly outperforms existing two-stage VPR methods in terms of computational efficiency. Code and models are available at https://github.com/chenshunpeng/FoL

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
AmsterTimeFoLRecall@170.1Unverified
AmsterTimeFoL-globalRecall@164.6Unverified
EynshamFoL-globalRecall@191.7Unverified
EynshamFoLRecall@192.4Unverified
Mapillary testFoL-globalRecall@178.7Unverified
Mapillary testFoLRecall@180Unverified
Mapillary valFoLRecall@193.5Unverified
Mapillary valFoL-globalRecall@193.1Unverified
NordlandFoL-globalRecall@187.8Unverified
NordlandFoLRecall@192.6Unverified
Nordland* (2760 queries)FoL-globalRecall@178.3Unverified
Nordland* (2760 queries)FoLRecall@185.5Unverified
Pittsburgh-250k-testFoLRecall@197Unverified
Pittsburgh-250k-testFoL-globalRecall@196.5Unverified
Pittsburgh-30k-testFoLRecall@194.5Unverified
Pittsburgh-30k-testFoL-globalRecall@193.9Unverified
SF-XL NightFoL-globalRecall@153.4Unverified
SF-XL NightFoLRecall@160.5Unverified
SF-XL OcclusionFoL-globalRecall@151.3Unverified
SF-XL OcclusionFoLRecall@161.8Unverified
SPEDFoLRecall@191.8Unverified
SPEDFoL-globalRecall@192.1Unverified
St LuciaFoLRecall@199.9Unverified
St LuciaFoL-globalRecall@199.9Unverified
SVOXFoLRecall@198.9Unverified
SVOXFoL-globalRecall@198.4Unverified
SVOX NightFoLRecall@198.8Unverified
SVOX NightFoL-globalRecall@198.3Unverified
SVOX-OvercastFoLRecall@198.2Unverified
SVOX-OvercastFoL-globalRecall@197.9Unverified
SVOX-RainFoL-globalRecall@196.5Unverified
SVOX-RainFoLRecall@198.2Unverified
SVOX-SnowFoL-globalRecall@199.1Unverified
SVOX-SnowFoLRecall@199.3Unverified
SVOX SunFoLRecall@198.8Unverified
SVOX SunFoL- globalRecall@198.1Unverified
Tokyo247FoL-globalRecall@196.2Unverified
Tokyo247FoLRecall@198.4Unverified

Reproductions