SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 67016725 of 8378 papers

TitleStatusHype
Delexicalized Paraphrase Generation0
Localization of Malaria Parasites and White Blood Cells in Thick Blood Smears0
FenceBox: A Platform for Defeating Adversarial Examples with Data Augmentation TechniquesCode0
Multi-Label Contrastive Learning for Abstract Visual ReasoningCode0
Intervention Design for Effective Sim2Real TransferCode0
How Robust are Randomized Smoothing based Defenses to Data Poisoning?0
A Self-Supervised Feature Map Augmentation (FMA) Loss and Combined Augmentations Finetuning to Efficiently Improve the Robustness of CNNs0
Chinese Grammatical Error Correction Based on Hybrid Models with Data Augmentation0
SkoltechNLP at SemEval-2020 Task 11: Exploring Unsupervised Text Augmentation for Propaganda Detection0
Denoising Pre-Training and Data Augmentation Strategies for Enhanced RDF Verbalization with Transformers0
Augmenting NLP models using Latent Feature Interpolations0
SMM4H Shared Task 2020 - A Hybrid Pipeline for Identifying Prescription Drug Abuse from Twitter: Machine Learning, Deep Learning, and Post-Processing0
Deep Subspace Clustering with Data Augmentation0
TMU Japanese-English Multimodal Machine Translation System for WAT 20200
BLCU-NLP at SemEval-2020 Task 5: Data Augmentation for Efficient Counterfactual Detecting0
Post-training Iterative Hierarchical Data Augmentation for Deep Networks0
Medication Mention Detection in Tweets Using ELECTRA Transformers and Decision Trees0
PhraseOut: A Code Mixed Data Augmentation Method for MultilingualNeural Machine Tranlsation0
XSYSIGMA at SemEval-2020 Task 7: Method for Predicting Headlines' Humor Based on Auxiliary Sentences with EI-BERT0
Improving Grammatical Error Correction with Data Augmentation by Editing Latent Representation0
Scalable Cross-lingual Treebank Synthesis for Improved Production Dependency Parsers0
Heterogeneous Recycle Generation for Chinese Grammatical Error Correction0
Parallel resources for Tunisian Arabic Dialect Translation0
Unifying Input and Output Smoothing in Neural Machine Translation0
The ADAPT Centre’s Neural MT Systems for the WAT 2020 Document-Level Translation Task0
Show:102550
← PrevPage 269 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified