SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 9761000 of 8378 papers

TitleStatusHype
Bootstrap Your Object Detector via Mixed TrainingCode1
VORTEX: Physics-Driven Data Augmentations Using Consistency Training for Robust Accelerated MRI ReconstructionCode1
Simple Conversational Data Augmentation for Semi-supervised Abstractive Dialogue SummarizationCode1
HypMix: Hyperbolic Interpolative Data AugmentationCode1
AESOP: Paraphrase Generation with Adaptive Syntactic ControlCode1
Towards the Generalization of Contrastive Self-Supervised LearningCode1
Efficiently Modeling Long Sequences with Structured State SpacesCode1
OMASGAN: Out-of-Distribution Minimum Anomaly Score GAN for Sample Generation on the BoundaryCode1
Robust Contrastive Learning Using Negative Samples with Diminished SemanticsCode1
How Important is Importance Sampling for Deep Budgeted Training?Code1
Identifying the key components in ResNet-50 for diabetic retinopathy grading from fundus images: a systematic investigationCode1
AugMax: Adversarial Composition of Random Augmentations for Robust TrainingCode1
Controllable Data Augmentation Through Deep RelightingCode1
Contextual Similarity Aggregation with Self-attention for Visual Re-rankingCode1
A Probabilistic Framework for Knowledge Graph Data AugmentationCode1
Contrastive Neural Processes for Self-Supervised LearningCode1
RCT: Random Consistency Training for Semi-supervised Sound Event DetectionCode1
NOD: Taking a Closer Look at Detection under Extreme Low-Light Conditions with Night Object Detection DatasetCode1
Improving Model Generalization by Agreement of Learned Representations from Data AugmentationCode1
Permutation invariant graph-to-sequence model for template-free retrosynthesis and reaction predictionCode1
Virtual Augmentation Supported Contrastive Learning of Sentence RepresentationsCode1
An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-trained Language ModelsCode1
MaGNET: Uniform Sampling from Deep Generative Network Manifolds Without RetrainingCode1
SGD-X: A Benchmark for Robust Generalization in Schema-Guided Dialogue SystemsCode1
Style-based quantum generative adversarial networks for Monte Carlo eventsCode1
Show:102550
← PrevPage 40 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified