SOTAVerified

Feature Upsampling

Deep features are a cornerstone of computer vision research, capturing image semantics and enabling the community to solve downstream tasks even in the zero- or few-shot regime. However, these features often lack the spatial resolution to directly perform dense prediction tasks like segmentation and depth prediction because models aggressively pool information over large areas. Feature Upsampling aims to recover this missing spatial resolution without impacting the space of the original deep features.

Papers

Showing 1120 of 25 papers

TitleStatusHype
SAPA: Similarity-Aware Point Affiliation for Feature UpsamplingCode1
FADE: Fusing the Assets of Decoder and Encoder for Task-Agnostic UpsamplingCode1
Local and Global GANs with Semantic-Aware Upsampling for Image GenerationCode1
Deep ViT Features as Dense Visual DescriptorsCode1
Joint Denoising and Demosaicking with Green Channel Prior for Real-world Burst ImagesCode1
Deep Image PriorCode1
Benchmarking Feature Upsampling Methods for Vision Foundation Models using Interactive SegmentationCode0
LDA-AQU: Adaptive Query-guided Upsampling via Local Deformable AttentionCode0
A Refreshed Similarity-based Upsampler for Direct High-Ratio Feature UpsamplingCode0
LiFT: A Surprisingly Simple Lightweight Feature Transform for Dense ViT DescriptorsCode0
Show:102550
← PrevPage 2 of 3Next →

No leaderboard results yet.