SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 11761200 of 8378 papers

TitleStatusHype
Improving Generalization by Controlling Label-Noise Information in Neural Network WeightsCode1
Contextual Similarity Aggregation with Self-attention for Visual Re-rankingCode1
Continual Few-shot Relation Learning via Embedding Space Regularization and Data AugmentationCode1
An Empirical Study of CLIP for Text-based Person SearchCode1
Contrastive Code Representation LearningCode1
ContrastCAD: Contrastive Learning-based Representation Learning for Computer-Aided Design ModelsCode1
Contrast and Classify: Training Robust VQA ModelsCode1
Contrastive Learning for Knowledge TracingCode1
Contrastive learning of global and local features for medical image segmentation with limited annotationsCode1
An Empirical Study on Channel Effects for Synthetic Voice Spoofing Countermeasure SystemsCode1
Contrastive Learning for Many-to-many Multilingual Neural Machine TranslationCode1
Contrastive Learning of User Behavior Sequence for Context-Aware Document RankingCode1
Contrastive Learning with Large Memory Bank and Negative Embedding Subtraction for Accurate Copy DetectionCode1
Incorporating External Knowledge through Pre-training for Natural Language to Code GenerationCode1
Abstract Meaning Representation-Based Logic-Driven Data Augmentation for Logical ReasoningCode1
A Shapelet-based Framework for Unsupervised Multivariate Time Series Representation LearningCode1
Contrastive Representation Learning for Gaze EstimationCode1
BET: A Backtranslation Approach for Easy Data Augmentation in Transformer-based Paraphrase Identification ContextCode1
Control, Generate, Augment: A Scalable Framework for Multi-Attribute Text GenerationCode1
Few-Shot Text Classification with Triplet Networks, Data Augmentation, and Curriculum LearningCode1
Injecting Numerical Reasoning Skills into Language ModelsCode1
Better plain ViT baselines for ImageNet-1kCode1
Controllable Data Augmentation Through Deep RelightingCode1
Better Robustness by More Coverage: Adversarial Training with Mixup Augmentation for Robust Fine-tuningCode1
Fine-Grained Egocentric Hand-Object Segmentation: Dataset, Model, and ApplicationsCode1
Show:102550
← PrevPage 48 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified