Shift-tolerant Perceptual Similarity Metric
Abhijay Ghildyal, Feng Liu
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/abhijay9/shifttolerant-lpipsOfficialIn paperpytorch★ 42
- github.com/farhadsh1992/stlpips_metric_tflitetf★ 0
Abstract
Existing perceptual similarity metrics assume an image and its reference are well aligned. As a result, these metrics are often sensitive to a small alignment error that is imperceptible to the human eyes. This paper studies the effect of small misalignment, specifically a small shift between the input and reference image, on existing metrics, and accordingly develops a shift-tolerant similarity metric. This paper builds upon LPIPS, a widely used learned perceptual similarity metric, and explores architectural design considerations to make it robust against imperceptible misalignment. Specifically, we study a wide spectrum of neural network elements, such as anti-aliasing filtering, pooling, striding, padding, and skip connection, and discuss their roles in making a robust metric. Based on our studies, we develop a new deep neural network-based perceptual similarity metric. Our experiments show that our metric is tolerant to imperceptible shifts while being consistent with the human similarity judgment.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| MSU SR-QA Dataset | ST-LPIPS (VGG) | SROCC | 0.57 | — | Unverified |
| MSU SR-QA Dataset | ST-LPIPS (Alex) | SROCC | 0.53 | — | Unverified |