SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 48014825 of 8378 papers

TitleStatusHype
Rock Classification Based on Residual Networks0
RocketQA: An Optimized Training Approach to Dense Passage Retrieval for Open-Domain Question Answering0
RocketQAv2: A Joint Training Method for Dense Passage Retrieval and Passage Re-ranking0
RoCNet: 3D Robust Registration of Point-Clouds using Deep Learning0
RoCoDA: Counterfactual Data Augmentation for Data-Efficient Robot Learning from Demonstrations0
RoFIR: Robust Fisheye Image Rectification Framework Impervious to Optical Center Deviation0
ROI-10D: Monocular Lifting of 2D Detection to 6D Pose and Metric Shape0
Role of Mixup in Topological Persistence Based Knowledge Distillation for Wearable Sensor Data0
Rollable Latent Space for Azimuth Invariant SAR Target Recognition0
Rolling the dice for better deep learning performance: A study of randomness techniques in deep neural networks0
ROMUL: Scale Adaptative Population Based Training0
RoPDA: Robust Prompt-based Data Augmentation for Low-Resource Named Entity Recognition0
Rotating spiders and reflecting dogs: a class conditional approach to learning data augmentation distributions0
Rotational augmentation techniques: a new perspective on ensemble learning for image classification0
Rotation Invariance and Extensive Data Augmentation: a strategy for the Mitosis Domain Generalization (MIDOG) Challenge0
Rotation-Invariant Local-to-Global Representation Learning for 3D Point Cloud0
RotEqNet: Rotation-Equivariant Network for Fluid Systems with Symmetric High-Order Tensors0
RRM: Robust Reward Model Training Mitigates Reward Hacking0
Rubric-Specific Approach to Automated Essay Scoring with Augmentation Training0
Rumour detection using graph neural network and oversampling in benchmark Twitter dataset0
RV-Syn: Rational and Verifiable Mathematical Reasoning Data Synthesis based on Structured Function Library0
S2M: Converting Single-Turn to Multi-Turn Datasets for Conversational Question Answering0
S3Aug: Segmentation, Sampling, and Shift for Action Recognition0
S^4M: Boosting Semi-Supervised Instance Segmentation with SAM0
S4RL: Surprisingly Simple Self-Supervision for Offline Reinforcement Learning0
Show:102550
← PrevPage 193 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified