SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 71767200 of 8378 papers

TitleStatusHype
VirAAL: Virtual Adversarial Active Learning For NLUCode0
You Do Not Need More Data: Improving End-To-End Speech Recognition by Text-To-Speech Data Augmentation0
Proxy Experience Replay: Federated Distillation for Distributed Reinforcement Learning0
ODVICE: An Ontology-Driven Visual Analytic Tool for Interactive Cohort Extraction0
Towards Robustifying NLI Models Against Lexical Dataset BiasesCode0
Deep Learning on Point Clouds for False Positive Reduction at Nodule Detection in Chest CT Scans0
Data Augmentation via Mixed Class Interpolation using Cycle-Consistent Generative Adversarial Networks Applied to Cross-Domain Imagery0
Data Augmentation for Hypernymy DetectionCode0
Correcting the Autocorrect: Context-Aware Typographical Error Correction via Training Data Augmentation0
Self-Training with Improved Regularization for Sample-Efficient Chest X-Ray Classification0
A Causal View on Robustness of Neural Networks0
A Comprehensive Survey of Grammar Error Correction0
Improving Non-autoregressive Neural Machine Translation with Monolingual Data0
Neural Symbolic Reader: Scalable Integration of Distributed and Symbolic Representations for Reading Comprehension0
Augmenting Small Data to Classify Contextualized Dialogue Acts for Exploratory Visualization0
When is Multi-task Learning Beneficial for Low-Resource Noisy Code-switched User-generated Algerian Texts?0
Stigma Annotation Scheme and Stigmatized Language Detection in Health-Care Discussions on Social Media0
On the Benefits of Invariance in Neural Networks0
Multilingual Neural Machine Translation involving Indian Languages0
Data Augmentation using Machine Translation for Fake News Detection in the Urdu Language0
Getting More Data for Low-resource Morphological Inflection: Language Models and Data Augmentation0
Multiword Expression aware Neural Machine Translation0
Conditional Augmentation for Aspect Term Extraction via Masked Sequence-to-Sequence Generation0
Does Data Augmentation Improve Generalization in NLP?0
Representations of Syntax [MASK] Useful: Effects of Constituency and Dependency Structure in Recursive LSTMsCode0
Show:102550
← PrevPage 288 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified