SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 58515900 of 8378 papers

TitleStatusHype
VOCABULARY-INFORMED VISUAL FEATURE AUGMENTATION FOR ONE-SHOT LEARNING0
Voice Conversion Can Improve ASR in Very Low-Resource Settings0
VoLux-GAN: A Generative Model for 3D Face Synthesis with HDRI Relighting0
VoronoiPatches: Evaluating A New Data Augmentation Method0
VPFNet: Improving 3D Object Detection with Virtual Point based LiDAR and Stereo Data Fusion0
WADER at SemEval-2023 Task 9: A Weak-labelling framework for Data augmentation in tExt Regression Tasks0
Wafer Map Defect Classification Using Autoencoder-Based Data Augmentation and Convolutional Neural Network0
Warping Resilient Scalable Anomaly Detection in Time Series0
WarriorCoder: Learning from Expert Battles to Augment Code Large Language Models0
Wasserstein Diffusion Tikhonov Regularization0
WatchGuardian: Enabling User-Defined Personalized Just-in-Time Intervention on Smartwatch0
Water Bottle Defect Detection System Using Convolutional Neural Network0
Wav2Vec-Aug: Improved self-supervised training with limited data0
Wav2vec-Switch: Contrastive Learning from Original-noisy Speech Pairs for Robust Speech Recognition0
Wavelet leader based formalism to compute multifractal features for classifying lung nodules in X-ray images0
Wavesplit: End-to-End Speech Separation by Speaker Clustering0
Weakly Supervised Data Augmentation Through Prompting for Dialogue Understanding0
Weakly Supervised Temporal Sentence Grounding With Uncertainty-Guided Self-Training0
Weakly supervised training of deep convolutional neural networks for overhead pedestrian localization in depth fields0
Weather Classification: A new multi-class dataset, data augmentation approach and comprehensive evaluations of Convolutional Neural Networks0
WeatherFormer: Empowering Global Numerical Weather Forecasting with Space-Time Transformer0
WeldMon: A Cost-effective Ultrasonic Welding Machine Condition Monitoring System0
WeMix: How to Better Utilize Data Augmentation0
WERank: Towards Rank Degradation Prevention for Self-Supervised Learning Using Weight Regularization0
WeSinger: Data-augmented Singing Voice Synthesis with Auxiliary Losses0
What Affects Learned Equivariance in Deep Image Recognition Models?0
What are effective labels for augmented data? Improving robustness with AutoLabel0
What Are Effective Labels for Augmented Data? Improving Calibration and Robustness with AutoLabel0
What Do Adversarially trained Neural Networks Focus: A Fourier Domain-based Study0
What do we learn from a large-scale study of pre-trained visual representations in sim and real environments?0
What Do You Need for Diverse Trajectory Stitching in Diffusion Planning?0
What Happened to My Dog in That Network: Unraveling Top-down Generators in Convolutional Neural Networks0
What is Holding Back Convnets for Detection?0
What makes a good data augmentation for few-shot unsupervised image anomaly detection?0
What Makes Better Augmentation Strategies? Augment Difficult but Not too Different0
What Makes for Good Views for Contrastive Learning?0
What Makes for Robust Multi-Modal Models in the Face of Missing Modalities?0
What Matters for Active Texture Recognition With Vision-Based Tactile Sensors0
What's All the FUSS About Free Universal Sound Separation Data?0
When and How Mixup Improves Calibration0
When Chosen Wisely, More Data Is What You Need: A Universal Sample-Efficient Strategy For Data Augmentation0
When Covariate-shifted Data Augmentation Increases Test Error And How to Fix It0
Does Data Augmentation Improve Generalization in NLP?0
When Does Re-initialization Work?0
When is Multi-task Learning Beneficial for Low-Resource Noisy Code-switched User-generated Algerian Texts?0
WHERE and WHICH: Iterative Debate for Biomedical Synthetic Data Augmentation0
Where is the bottleneck in long-tailed classification?0
Where is the disease? Semi-supervised pseudo-normality synthesis from an abnormal image0
Where Should I Spend My FLOPS? Efficiency Evaluations of Visual Pre-training Methods0
Which Student is Best? A Comprehensive Knowledge Distillation Exam for Task-Specific BERT Models0
Show:102550
← PrevPage 118 of 168Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified