SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 56765700 of 8378 papers

TitleStatusHype
UAVTwin: Neural Digital Twins for UAVs using Gaussian Splatting0
UDRN: Unified Dimensional Reduction Neural Network for Feature Selection and Feature Projection0
UFRC: A Unified Framework for Reliable COVID-19 Detection on Crowdsourced Cough Audio0
UI Layers Group Detector: Grouping UI Layers via Text Fusion and Box Attention0
Ulcerative Colitis Mayo Endoscopic Scoring Classification with Active Learning and Generative Data Augmentation0
Ultrafast-and-Ultralight ConvNet-Based Intelligent Monitoring System for Diagnosing Early-Stage Mpox Anytime and Anywhere0
UMass PCL at SemEval-2022 Task 4: Pre-trained Language Model Ensembles for Detecting Patronizing and Condescending Language0
Unbiased Faster R-CNN for Single-source Domain Generalized Object Detection0
Uncertainty Estimation and Out-of-Distribution Detection for Deep Learning-Based Image Reconstruction using the Local Lipschitz0
Unconstrained Road Marking Recognition with Generative Adversarial Networks0
Undersensitivity in Neural Reading Comprehension0
Understanding Adversarial Robustness Through Loss Landscape Geometries0
Gradient-Leaks: Understanding and Controlling Deanonymization in Federated Learning0
Understanding and Improving Robustness of Vision Transformers through Patch-based Negative Augmentation0
Understanding and Mitigating Memorization in Diffusion Models for Tabular Data0
Understanding and Mitigating the Bias Inheritance in LLM-based Data Augmentation on Downstream Tasks0
Understanding Augmentation-based Self-Supervised Representation Learning via RKHS Approximation and Regression0
On the Calibration of Multilingual Question Answering LLMs0
Understanding Community Bias Amplification in Graph Representation Learning0
Understanding data augmentation for classification: when to warp?0
Understanding Data Augmentation from a Robustness Perspective0
Understanding Data Augmentation in Neural Machine Translation: Two Perspectives towards Generalization0
PAC-Chernoff Bounds: Understanding Generalization in the Interpolation Regime0
Understanding Learning Invariance in Deep Linear Networks0
Understanding Masked Autoencoders From a Local Contrastive Perspective0
Show:102550
← PrevPage 228 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified