SOTAVerified

TransAdapter: Vision Transformer for Feature-Centric Unsupervised Domain Adaptation

2024-12-05Code Available1· sign in to hype

A. Enes Doruk, Erhan Oztop, Hasan F. Ates

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Unsupervised Domain Adaptation (UDA) aims to utilize labeled data from a source domain to solve tasks in an unlabeled target domain, often hindered by significant domain gaps. Traditional CNN-based methods struggle to fully capture complex domain relationships, motivating the shift to vision transformers like the Swin Transformer, which excel in modeling both local and global dependencies. In this work, we propose a novel UDA approach leveraging the Swin Transformer with three key modules. A Graph Domain Discriminator enhances domain alignment by capturing inter-pixel correlations through graph convolutions and entropy-based attention differentiation. An Adaptive Double Attention module combines Windows and Shifted Windows attention with dynamic reweighting to align long-range and local features effectively. Finally, a Cross-Feature Transform modifies Swin Transformer blocks to improve generalization across domains. Extensive benchmarks confirm the state-of-the-art performance of our versatile method, which requires no task-specific alignment modules, establishing its adaptability to diverse applications.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
DomainNetTransadapterAccuracy53.7Unverified
Office-HomeTransAdapter-BAccuracy89.4Unverified
VisDA-2017TransAdapterAccuracy91.2Unverified
VisDA2017TransAdapterAccuracy91.2Unverified

Reproductions