SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 62766300 of 8378 papers

TitleStatusHype
NOSE Augment: Fast and Effective Data Augmentation Without Searching0
Logit As Auxiliary Weak-supervision for More Reliable and Accurate Prediction0
Generalization and Stability of GANs: A theory and promise from data augmentation0
Neural Networks Preserve Invertibility Across Iterations: A Possible Source of Implicit Data Augmentation0
XLA: A Robust Unsupervised Data Augmentation Framework for Cross-Lingual NLP0
Redefining Self-Normalization Property0
Recurrent Neural Network Architecture based on Dynamic Systems Theory for Data Driven Modelling of Complex Physical Systems0
DJMix: Unsupervised Task-agnostic Augmentation for Improving Robustness0
Tradeoffs in Data Augmentation: An Empirical Study0
On the Consistency Loss for Leveraging Augmented Data to Learn Robust and Invariant Representations0
Switching-Aligned-Words Data Augmentation for Neural Machine Translation0
Disentangling style and content for low resource video domain adaptation: a case study on keystroke inference attacks0
Transferable Unsupervised Robust Representation Learning0
Generative Auto-Encoder: Non-adversarial Controllable Synthesis with Disentangled Exploration0
PriorityCut: Occlusion-aware Regularization for Image Animation0
A Rigorous Evaluation of Real-World Distribution Shifts0
The Importance of Importance Sampling for Deep Budgeted Training0
Driving through the Lens: Improving Generalization of Learning-based Steering using Simulated Adversarial Examples0
Data augmentation for deep learning based accelerated MRI reconstruction0
What are effective labels for augmented data? Improving robustness with AutoLabel0
Cut out the annotator, keep the cutout: better segmentation with weak supervision0
Improving Generalizability of Protein Sequence Models via Data Augmentations0
ROMUL: Scale Adaptative Population Based Training0
FastIF: Scalable Influence Functions for Efficient Model Interpretation and DebuggingCode0
Better Robustness by More Coverage: Adversarial Training with Mixup Augmentation for Robust Fine-tuningCode1
Show:102550
← PrevPage 252 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified