SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 651700 of 8378 papers

TitleStatusHype
CACTI: A Framework for Scalable Multi-Task Multi-Scene Visual Imitation LearningCode1
On Pre-Training for Visuo-Motor Control: Revisiting a Learning-from-Scratch BaselineCode1
Robust and Explainable Identification of Logical Fallacies in Natural Language ArgumentsCode1
Towards Scale Balanced 6-DoF Grasp Detection in Cluttered ScenesCode1
X-Paste: Revisiting Scalable Copy-Paste for Instance Segmentation using CLIP and StableDiffusionCode1
ObjectStitch: Generative Object CompositingCode1
CL4CTR: A Contrastive Learning Framework for CTR PredictionCode1
Improving Commonsense in Vision-Language Models via Knowledge Graph RiddlesCode1
RGB no more: Minimally-decoded JPEG Vision TransformersCode1
Rethinking Data Augmentation for Single-source Domain Generalization in Medical Image SegmentationCode1
Towards Good Practices for Missing Modality Robust Action RecognitionCode1
Pose-disentangled Contrastive Learning for Self-supervised Facial RepresentationCode1
Mitigating and Evaluating Static Bias of Action Representations in the Background and the ForegroundCode1
Mitigating Data Sparsity for Short Text Topic Modeling by Topic-Semantic Contrastive LearningCode1
Join the High Accuracy Club on ImageNet with A Binary Neural Network TicketCode1
Breaking the Representation Bottleneck of Chinese Characters: Neural Machine Translation with Stroke Sequence ModelingCode1
RoentGen: Vision-Language Foundation Model for Chest X-ray GenerationCode1
ModelDiff: A Framework for Comparing Learning AlgorithmsCode1
Multimodal Data Augmentation for Visual-Infrared Person ReID with Corrupted DataCode1
Fed-TDA: Federated Tabular Data Augmentation on Non-IID DataCode1
EVNet: An Explainable Deep Network for Dimension ReductionCode1
RIC-CNN: Rotation-Invariant Coordinate Convolutional Neural NetworkCode1
Background-Mixed Augmentation for Weakly Supervised Change DetectionCode1
SeeABLE: Soft Discrepancies and Bounded Contrastive Learning for Exposing DeepfakesCode1
GENIUS: Sketch-based Language Model Pre-training via Extreme and Selective Masking for Text Generation and AugmentationCode1
A Unified Multimodal De- and Re-coupling Framework for RGB-D Motion RecognitionCode1
CST5: Data Augmentation for Code-Switched Semantic ParsingCode1
MDFlow: Unsupervised Optical Flow Learning by Reliable Mutual Knowledge DistillationCode1
DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question AnsweringCode1
IRNet: Iterative Refinement Network for Noisy Partial Label LearningCode1
Soft Augmentation for Image ClassificationCode1
GOOD-D: On Unsupervised Graph Out-Of-Distribution DetectionCode1
Unleashing the Power of Graph Data Augmentation on Covariate Distribution ShiftCode1
Effective Audio Classification Network Based on Paired Inverse Pyramid Structure and Dense MLP BlockCode1
Conditional Generative Models for Simulation of EMG During Naturalistic MovementsCode1
Self-supervised Character-to-Character Distillation for Text RecognitionCode1
Why is Winoground Hard? Investigating Failures in Visuolinguistic CompositionalityCode1
Rethinking and Improving Robustness of Convolutional Neural Networks: a Shapley Value-based Approach in Frequency DomainCode1
Towards Better Out-of-Distribution Generalization of Neural Algorithmic Reasoning TasksCode1
Differentiable Data Augmentation for Contrastive Sentence Representation LearningCode1
Speaker Representation Learning via Contrastive Loss with Maximal Speaker SeparabilityCode1
RoChBert: Towards Robust BERT Fine-tuning for ChineseCode1
U-Net-based Models for Skin Lesion Segmentation: More Attention and AugmentationCode1
Domain Adaptive Object Detection for Autonomous Driving under Foggy WeatherCode1
ADLight: A Universal Approach of Traffic Signal Control with Augmented Data Using Reinforcement LearningCode1
Contrastive Representation Learning for Gaze EstimationCode1
Rethinking Rotation in Self-Supervised Contrastive Learning: Adaptive Positive or Negative Data AugmentationCode1
Neural Eigenfunctions Are Structured Representation LearnersCode1
Exploring Representation-Level Augmentation for Code SearchCode1
RMBench: Benchmarking Deep Reinforcement Learning for Robotic Manipulator ControlCode1
Show:102550
← PrevPage 14 of 168Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified