SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 10011025 of 8378 papers

TitleStatusHype
UniSpeech-SAT: Universal Speech Representation Learning with Speaker Aware Pre-TrainingCode1
Point Cloud Augmentation with Weighted Local TransformationsCode1
Semi-Supervised Semantic Segmentation via Adaptive Equalization LearningCode1
Learning 3D Representations of Molecular Chirality with Invariance to Bond RotationsCode1
Towards Accurate Cross-Domain In-Bed Human Pose EstimationCode1
StrengthNet: Deep Learning-based Emotion Strength Assessment for Emotional Speech SynthesisCode1
FilterAugment: An Acoustic Environmental Data Augmentation MethodCode1
Deep Transfer Learning for Land Use and Land Cover Classification: A Comparative StudyCode1
Shallow Features Guide Unsupervised Domain Adaptation for Semantic Segmentation at Class BoundariesCode1
Noisy Feature MixupCode1
Self-Supervised Generative Style Transfer for One-Shot Medical Image SegmentationCode1
Data Augmentation Approaches in Natural Language Processing: A SurveyCode1
Transfer Learning U-Net Deep Learning for Lung Ultrasound SegmentationCode1
WaveBeat: End-to-end beat and downbeat tracking in the time domainCode1
GenCo: Generative Co-training for Generative Adversarial Networks with Limited DataCode1
Offline Reinforcement Learning with Reverse Model-based ImaginationCode1
ResNet strikes back: An improved training procedure in timmCode1
Instance Segmentation Challenge Track Technical Report, VIPriors Workshop at ICCV 2021: Task-Specific Copy-Paste Data Augmentation Method for Instance SegmentationCode1
CALDA: Improving Multi-Source Time Series Domain Adaptation with Contrastive Adversarial LearningCode1
Natural Synthetic Anomalies for Self-Supervised Anomaly Detection and LocalizationCode1
Target-Side Data Augmentation for Sequence GenerationCode1
AdaAug: Learning Class- and Instance-adaptive Data Augmentation PoliciesCode1
NASViT: Neural Architecture Search for Efficient Vision Transformers with Gradient Conflict aware Supernet TrainingCode1
Stochastic Training is Not Necessary for GeneralizationCode1
Excavating the Potential Capacity of Self-Supervised Monocular Depth EstimationCode1
Show:102550
← PrevPage 41 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified