SOTAVerified

Metric Learning

The goal of Metric Learning is to learn a representation function that maps objects into an embedded space. The distance in the embedded space should preserve the objects’ similarity — similar objects get close and dissimilar objects get far away. Various loss functions have been developed for Metric Learning. For example, the contrastive loss guides the objects from the same class to be mapped to the same point and those from different classes to be mapped to different points whose distances are larger than a margin. Triplet loss is also popular, which requires the distance between the anchor sample and the positive sample to be smaller than the distance between the anchor sample and the negative sample.

Source: Road Network Metric Learning for Estimated Time of Arrival

Papers

Showing 211220 of 1648 papers

TitleStatusHype
OpenGAN: Open Set Generative Adversarial NetworksCode1
Optimizing Bi-Encoder for Named Entity Recognition via Contrastive LearningCode1
PANet: Few-Shot Image Semantic Segmentation with Prototype AlignmentCode1
Patch-wise Deep Metric Learning for Unsupervised Low-Dose CT DenoisingCode1
Characterizing Generalization under Out-Of-Distribution Shifts in Deep Metric LearningCode1
A Comparison of Metric Learning Loss Functions for End-To-End Speaker VerificationCode1
Circle Loss: A Unified Perspective of Pair Similarity OptimizationCode1
PieNet: Personalized Image Enhancement NetworkCode1
Probabilistic Embeddings RevisitedCode1
Isotropy Maximization Loss and Entropic Score: Accurate, Fast, Efficient, Scalable, and Turnkey Neural Networks Out-of-Distribution Detection Based on The Principle of Maximum EntropyCode1
Show:102550
← PrevPage 22 of 165Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Unicom+ViT-L@336pxR@198.2Unverified
2Hyp-DINO 8x8R@192.8Unverified
3ResNet-50 + AVSLR@191.5Unverified
4NEDR@191.5Unverified
5ResNet-50 + Intra-Batch (ensemble of 5)R@191.5Unverified
6EfficientDML-VPTSP-G/512R@191.2Unverified
7CCL (ResNet-50)R@191.02Unverified
8ResNet50 + LanguageR@190.2Unverified
9ResNet-50 + MetrixR@189.6Unverified
10ResNet50 + S2SDR@189.5Unverified
#ModelMetricClaimedVerifiedStatus
1Unicom+ViT-L@336pxR@191.2Unverified
2STIRR@188.3Unverified
3Recall@k Surrogate Loss (ViT-B/16)R@188Unverified
4ViT-TripletR@186.5Unverified
5ROADMAP (DeiT-S)R@186Unverified
6Hyp-ViTR@185.9Unverified
7Hyp-DINOR@185.1Unverified
8Recall@k Surrogate Loss (ViT-B/32)R@185.1Unverified
9CCL (ResNet-50)R@183.1Unverified
10ROADMAP (ResNet-50)R@183.1Unverified
#ModelMetricClaimedVerifiedStatus
1Unicom+ViT-L@336pxR@190.1Unverified
2EfficientDML-VPTSP-G/512R@188.5Unverified
3Hyp-ViTR@185.6Unverified
4Hyp-DINOR@180.9Unverified
5NEDR@174.9Unverified
6CCL (ResNet-50)R@173.45Unverified
7ResNet-50 + AVSLR@171.9Unverified
8ResNet-50 + Intra-Batch ConnectionsR@171.8Unverified
9ResNet50 + LanguageR@171.4Unverified
10ResNet-50 + MetrixR@171.4Unverified
#ModelMetricClaimedVerifiedStatus
1Unicom+ViT-L@336pxR@196.7Unverified
2STIRR@195Unverified
3MGAR@194.3Unverified
4Hyp-ViTR@192.5Unverified
5Hyp-DINOR@192.4Unverified
6CCL (ResNet-50)R@192.31Unverified
7Gradient SurgeryR@192.21Unverified
8ResNet-50 + MetrixR@192.2Unverified
9EfficientDML-VPTSP-G/512R@192.1Unverified
10ViT-TripletR@192.1Unverified
#ModelMetricClaimedVerifiedStatus
1HAPPIERAverage-mAP43.8Unverified
2CSLAverage-mAP31Unverified
#ModelMetricClaimedVerifiedStatus
1HAPPIERAverage-mAP38Unverified
2CSLAverage-mAP28.7Unverified
#ModelMetricClaimedVerifiedStatus
1HAPPIERAverage-mAP37Unverified
2CSLAverage-mAP12.1Unverified