SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 44264450 of 8378 papers

TitleStatusHype
Balancing Data through Data Augmentation Improves the Generality of Transfer Learning for Diabetic Retinopathy Classification0
Audio Data Augmentation for Acoustic-to-articulatory Speech Inversion using Bidirectional Gated RNNs0
Investigating Lexical Replacements for Arabic-English Code-Switched Data Augmentation0
Augmentation-induced Consistency Regularization for Classification0
Leveraging QA Datasets to Improve Generative Data AugmentationCode0
Conditional set generation using Seq2seq models0
An Empirical Study on Distribution Shift Robustness From the Perspective of Pre-Training and Data Augmentation0
Does Your Model Classify Entities Reasonably? Diagnosing and Mitigating Spurious Correlations in Entity TypingCode0
Counterfactual Data Augmentation improves Factuality of Abstractive Summarization0
ReSmooth: Detecting and Utilizing OOD Samples when Training with Data AugmentationCode1
Medical Scientific Table-to-Text Generation with Human-in-the-Loop under the Data Sparsity Constraint0
Multi-Augmentation for Efficient Visual Representation Learning for Self-supervised Pre-trainingCode0
Robust 3D Object Detection in Cold Weather Conditions0
Deep Learning-based automated classification of Chinese Speech Sound Disorders0
EventMix: An Efficient Augmentation Strategy for Event-Based Data0
Highly Accurate FMRI ADHD Classification using time distributed multi modal 3D CNNsCode1
One-Pixel Shortcut: on the Learning Preference of Deep Neural NetworksCode1
Training Efficient CNNS: Tweaking the Nuts and Bolts of Neural Networks for Lighter, Faster and Robust ModelsCode0
Learning to Ignore Adversarial Attacks0
QASem Parsing: Text-to-text Modeling of QA-based SemanticsCode1
Data augmentation for efficient learning from parametric experts0
Sleep Posture One-Shot Learning Framework Using Kinematic Data Augmentation: In-Silico and In-Vivo Case Studies0
CNNs Avoid Curse of Dimensionality by Learning on Patches0
Self-mentoring: a new deep learning pipeline to train a self-supervised U-net for few-shot learning of bio-artificial capsule segmentation0
Producing Histopathology Phantom Images using Generative Adversarial Networks to improve Tumor Detection0
Show:102550
← PrevPage 178 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified