SOTAVerified

Out of Distribution (OOD) Detection

Out of Distribution (OOD) Detection is the task of detecting instances that do not belong to the distribution the classifier has been trained on. OOD data is often referred to as "unseen" data, as the model has not encountered it during training.

OOD detection is typically performed by training a model to distinguish between in-distribution (ID) data, which the model has seen during training, and OOD data, which it has not seen. This can be done using a variety of techniques, such as training a separate OOD detector, or modifying the model's architecture or loss function to make it more sensitive to OOD data.

Papers

Showing 261270 of 629 papers

TitleStatusHype
Effectiveness of Vision Language Models for Open-world Single Image Test Time Adaptation0
When and How Does In-Distribution Label Help Out-of-Distribution Detection?Code0
WeiPer: OOD Detection using Weight Perturbations of Class Projections0
Concept Matching with Agent for Out-of-Distribution DetectionCode0
Enhancing Near OOD Detection in Prompt Learning: Maximum Gains, Minimal Costs0
Dual-Adapter: Training-free Dual Adaptation for Few-shot Out-of-Distribution Detection0
Credal Wrapper of Model Averaging for Uncertainty Estimation in Classification0
A Unified Approach Towards Active Learning and Out-of-Distribution Detection0
Multi-Label Out-of-Distribution Detection with Spectral Normalized Joint EnergyCode0
Language-Enhanced Latent Representations for Out-of-Distribution Detection in Autonomous Driving0
Show:102550
← PrevPage 27 of 63Next →

No leaderboard results yet.