SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 43514375 of 8378 papers

TitleStatusHype
Masked Autoencoders are Robust Data AugmentorsCode1
Is Self-Supervised Learning More Robust Than Supervised Learning?0
Heterogeneous Face Recognition via Face Synthesis with Identity-Attribute Disentanglement0
Extreme Masking for Learning Instance and Distributed Visual RepresentationsCode1
I'm Me, We're Us, and I'm Us: Tri-directional Contrastive Learning on HypergraphsCode1
Beyond RGB: Scene-Property Synthesis with Neural Radiance Fields0
Learn2Augment: Learning to Composite Videos for Data Augmentation in Action Recognition0
BSM loss: A superior way in modeling aleatory uncertainty of fine_grained classification0
PointNeXt: Revisiting PointNet++ with Improved Training and Scaling StrategiesCode3
Metric Based Few-Shot Graph ClassificationCode1
On gradient descent training under data augmentation with on-line noisy copies0
Always Keep your Target in Mind: Studying Semantics and Improving Performance of Neural Lexical SubstitutionCode0
An efficient semi-supervised quality control system trained using physics-based MRI-artefact generators and adversarial training0
Marvolo: Programmatic Data Augmentation for Practical ML-Driven Malware Detection0
PP-OCRv3: More Attempts for the Improvement of Ultra Lightweight OCR System0
Mixed Graph Contrastive Network for Semi-Supervised Node Classification0
Global Mixup: Eliminating Ambiguity with Clustering0
Stacked unsupervised learning with a network architecture found by supervised meta-learning0
AugLoss: A Robust Augmentation-based Fine Tuning Methodology0
Toward Learning Robust and Invariant Representations with Alignment Regularization and Data AugmentationCode1
Monkeypox Image Data collectionCode1
Integrating Prior Knowledge in Contrastive Learning with KernelCode0
Adversarial Unlearning: Reducing Confidence Along Adversarial Directions0
YOLOv5s-GTB: light-weighted and improved YOLOv5s for bridge crack detection0
MaxStyle: Adversarial Style Composition for Robust Medical Image SegmentationCode1
Show:102550
← PrevPage 175 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified