SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 876900 of 8378 papers

TitleStatusHype
DiffBatt: A Diffusion Model for Battery Degradation Prediction and SynthesisCode1
Counterfactual MRI Data Augmentation using Conditional Denoising Diffusion Generative ModelsCode0
Provable Benefit of Cutout and CutMix for Feature Learning0
Generative forecasting of brain activity enhances Alzheimer's classification and interpretation0
DAVINCI: A Single-Stage Architecture for Constrained CAD Sketch InferenceCode1
First Place Solution to the ECCV 2024 ROAD++ Challenge @ ROAD++ Spatiotemporal Agent Detection 20240
SleepNetZero: Zero-Burden Zero-Shot Reliable Sleep Staging With Neural Networks Based on Ballistocardiograms0
First Place Solution to the ECCV 2024 ROAD++ Challenge @ ROAD++ Atomic Activity Recognition 20240
Does equivariance matter at scale?0
Saliency-Based diversity and fairness Metric and FaceKeepOriginalAugment: A Novel Approach for Enhancing Fairness and Diversity0
NetAurHPD: Network Auralization Hyperlink Prediction Model to Identify Metabolic Pathways from Metabolomics DataCode0
Guided Diffusion-based Counterfactual Augmentation for Robust Session-based Recommendation0
Attention Speaks Volumes: Localizing and Mitigating Bias in Language Models0
FairSkin: Fair Diffusion for Skin Disease Image Generation0
Data Generation for Hardware-Friendly Post-Training QuantizationCode3
Not All LLM-Generated Data Are Equal: Rethinking Data Weighting in Text Classification0
LinFormer: A Linear-based Lightweight Transformer Architecture For Time-Aware MIMO Channel Prediction0
Synthetica: Large Scale Synthetic Data for Robot Perception0
Scaling-based Data Augmentation for Generative Models and its Theoretical Extension0
Relation-based Counterfactual Data Augmentation and Contrastive Learning for Robustifying Natural Language Inference Models0
Mitigating Unauthorized Speech Synthesis for Voice ProtectionCode1
BongLLaMA: LLaMA for Bangla Language0
Unsupervised Panoptic Interpretation of Latent Spaces in GANs Using Space-Filling Vector QuantizationCode0
Guiding Through Complexity: What Makes Good Supervision for Hard Reasoning Tasks?Code0
SAFE setup for generative molecular design0
Show:102550
← PrevPage 36 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified