SOTAVerified

Metric Learning

The goal of Metric Learning is to learn a representation function that maps objects into an embedded space. The distance in the embedded space should preserve the objects’ similarity — similar objects get close and dissimilar objects get far away. Various loss functions have been developed for Metric Learning. For example, the contrastive loss guides the objects from the same class to be mapped to the same point and those from different classes to be mapped to different points whose distances are larger than a margin. Triplet loss is also popular, which requires the distance between the anchor sample and the positive sample to be smaller than the distance between the anchor sample and the negative sample.

Source: Road Network Metric Learning for Estimated Time of Arrival

Papers

Showing 14811490 of 1648 papers

TitleStatusHype
Task-Embedded Control Networks for Few-Shot Imitation LearningCode0
Local Metric Learning for Off-Policy Evaluation in Contextual Bandits with Continuous ActionsCode0
Long-Tailed Classification Based on Coarse-Grained Leading Forest and Multi-Center LossCode0
Distance-Ratio-Based Formulation for Metric LearningCode0
Noise-robust Graph Learning by Estimating and Leveraging Pairwise InteractionsCode0
LoOp: Looking for Optimal Hard Negative Embeddings for Deep Metric LearningCode0
Low-rank geometric mean metric learningCode0
Batch-Incremental Triplet Sampling for Training Triplet Networks Using Bayesian Updating TheoremCode0
M^2S-Net: Multi-Modal Similarity Metric Learning based Deep Convolutional Network for Answer SelectionCode0
Cross-View Kernel Similarity Metric Learning Using Pairwise Constraints for Person Re-identificationCode0
Show:102550
← PrevPage 149 of 165Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Unicom+ViT-L@336pxR@198.2Unverified
2Hyp-DINO 8x8R@192.8Unverified
3ResNet-50 + AVSLR@191.5Unverified
4NEDR@191.5Unverified
5ResNet-50 + Intra-Batch (ensemble of 5)R@191.5Unverified
6EfficientDML-VPTSP-G/512R@191.2Unverified
7CCL (ResNet-50)R@191.02Unverified
8ResNet50 + LanguageR@190.2Unverified
9ResNet-50 + MetrixR@189.6Unverified
10ResNet50 + S2SDR@189.5Unverified
#ModelMetricClaimedVerifiedStatus
1Unicom+ViT-L@336pxR@191.2Unverified
2STIRR@188.3Unverified
3Recall@k Surrogate Loss (ViT-B/16)R@188Unverified
4ViT-TripletR@186.5Unverified
5ROADMAP (DeiT-S)R@186Unverified
6Hyp-ViTR@185.9Unverified
7Hyp-DINOR@185.1Unverified
8Recall@k Surrogate Loss (ViT-B/32)R@185.1Unverified
9CCL (ResNet-50)R@183.1Unverified
10ROADMAP (ResNet-50)R@183.1Unverified
#ModelMetricClaimedVerifiedStatus
1Unicom+ViT-L@336pxR@190.1Unverified
2EfficientDML-VPTSP-G/512R@188.5Unverified
3Hyp-ViTR@185.6Unverified
4Hyp-DINOR@180.9Unverified
5NEDR@174.9Unverified
6CCL (ResNet-50)R@173.45Unverified
7ResNet-50 + AVSLR@171.9Unverified
8ResNet-50 + Intra-Batch ConnectionsR@171.8Unverified
9ResNet50 + LanguageR@171.4Unverified
10ResNet-50 + MetrixR@171.4Unverified
#ModelMetricClaimedVerifiedStatus
1Unicom+ViT-L@336pxR@196.7Unverified
2STIRR@195Unverified
3MGAR@194.3Unverified
4Hyp-ViTR@192.5Unverified
5Hyp-DINOR@192.4Unverified
6CCL (ResNet-50)R@192.31Unverified
7Gradient SurgeryR@192.21Unverified
8ResNet-50 + MetrixR@192.2Unverified
9EfficientDML-VPTSP-G/512R@192.1Unverified
10ViT-TripletR@192.1Unverified
#ModelMetricClaimedVerifiedStatus
1HAPPIERAverage-mAP43.8Unverified
2CSLAverage-mAP31Unverified
#ModelMetricClaimedVerifiedStatus
1HAPPIERAverage-mAP38Unverified
2CSLAverage-mAP28.7Unverified
#ModelMetricClaimedVerifiedStatus
1HAPPIERAverage-mAP37Unverified
2CSLAverage-mAP12.1Unverified