SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 12761300 of 8378 papers

TitleStatusHype
Graph Transformer for RecommendationCode1
GRLib: An Open-Source Hand Gesture Detection and Recognition Python LibraryCode1
Grounded Adaptation for Zero-shot Executable Semantic ParsingCode1
A Comprehensive Approach to Unsupervised Embedding Learning based on AND AlgorithmCode1
CLARA: Multilingual Contrastive Learning for Audio Representation AcquisitionCode1
AnnoCTR: A Dataset for Detecting and Linking Entities, Tactics, and Techniques in Cyber Threat ReportsCode1
Harmonic Networks: Deep Translation and Rotation EquivarianceCode1
MixRec: Heterogeneous Graph Collaborative FilteringCode1
HFL at SemEval-2022 Task 8: A Linguistics-inspired Regression Model with Data Augmentation for Multilingual News SimilarityCode1
Hierarchical Amortized Training for Memory-efficient High Resolution 3D GANCode1
Hierarchical Disentanglement-Alignment Network for Robust SAR Vehicle RecognitionCode1
Contrast and Classify: Training Robust VQA ModelsCode1
Highly Accurate FMRI ADHD Classification using time distributed multi modal 3D CNNsCode1
HINER: Neural Representation for Hyperspectral ImageCode1
Histogram Matching Augmentation for Domain Adaptation with Application to Multi-Centre, Multi-Vendor and Multi-Disease Cardiac Image SegmentationCode1
How Important is Importance Sampling for Deep Budgeted Training?Code1
How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language UnderstandingCode1
DiffAug: Enhance Unsupervised Contrastive Learning with Domain-Knowledge-Free Diffusion-based Data AugmentationCode1
BOOTPLACE: Bootstrapped Object Placement with Detection TransformersCode1
Causal Action Influence Aware Counterfactual Data AugmentationCode1
HybridAugment++: Unified Frequency Spectra Perturbations for Model RobustnessCode1
HyperTab: Hypernetwork Approach for Deep Learning on Small Tabular DatasetsCode1
Bootstrapping Relation Extractors using Syntactic Search by ExamplesCode1
ID-Blau: Image Deblurring by Implicit Diffusion-based reBLurring AUgmentationCode1
Contrastive Code Representation LearningCode1
Show:102550
← PrevPage 52 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified