SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 61266150 of 8378 papers

TitleStatusHype
Learning for Counterfactual Fairness from Observational Data0
Learning from Data-Rich Problems: A Case Study on Genetic Variant Calling0
Learning from Few Samples: A Survey0
Learning from Multiple Noisy Augmented Data Sets for Better Cross-Lingual Spoken Language Understanding0
Learning from One Continuous Video Stream0
Learning Gaussian Data Augmentation in Feature Space for One-shot Object Detection in Manga0
Learning Generalizable Models via Disentangling Spurious and Enhancing Potential Correlations0
Learning Human Perception Dynamics for Informative Robot Communication0
Learning Instance-Specific Adaptation for Cross-Domain Segmentation0
Learning Invariances using the Marginal Likelihood0
Learning Knowledge-Enhanced Contextual Language Representations for Domain Natural Language Understanding0
Learning linear modules in a dynamic network with missing node observations0
Learning Linear Symmetries in Data Using Moment Matching0
Learning Loss for Test-Time Augmentation0
Learning More with Less: Conditional PGGAN-based Data Augmentation for Brain Metastases Detection Using Highly-Rough Annotation on MR Images0
Learning More with Less: GAN-based Medical Image Augmentation0
Learning Neural Light Transport0
Learning Noise-Invariant Representations for Robust Speech Recognition0
Learning Noise-Invariant Representations for Robust Speech Recognition0
Learning Noise-Robust Stable Koopman Operator for Control with Hankel DMD0
Learning Nonlinear Brain Dynamics: van der Pol Meets LSTM0
Learning Object Placement by Inpainting for Compositional Data Augmentation0
Learning Online Visual Invariances for Novel Objects via Supervised and Self-Supervised Training0
Learning Physical Common Sense as Knowledge Graph Completion via BERT Data Augmentation and Constrained Tucker Factorization0
Learning Polynomial Problems with SL(2,R) Equivariance0
Show:102550
← PrevPage 246 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified