SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 50765100 of 8378 papers

TitleStatusHype
Causality-inspired Single-source Domain Generalization for Medical Image SegmentationCode1
Revisiting Contextual Toxicity Detection in Conversations0
Multi-Modality Microscopy Image Style Transfer for Nuclei Segmentation0
AutoDC: Automated data-centric processingCode1
Weight Pruning and Uncertainty in Radio Galaxy ClassificationCode0
Domain-Agnostic Clustering with Self-Distillation0
S-SimCSE: Sampled Sub-networks for Contrastive Learning of Sentence Embedding0
Variance Reduction in Deep Learning: More Momentum is All You Need0
Using mixup as regularization and tuning hyper-parameters for ResNetsCode0
Broad Adversarial Training with Data Augmentation in the Output Space0
MUM : Mix Image Tiles and UnMix Feature Tiles for Semi-Supervised Object DetectionCode1
Why Do Self-Supervised Models Transfer? Investigating the Impact of Invariance on Downstream TasksCode1
Video Content Swapping Using GAN0
Unsupervised Domain Adaptation for RF-based Gesture Recognition0
Simulated LiDAR Repositioning: a novel point cloud data augmentation method0
Toxicity Detection can be Sensitive to the Conversational Context0
Improved Prosodic Clustering for Multispeaker and Speaker-independent Phoneme-level Prosody Control0
An Analysis of the Influence of Transfer Learning When Measuring the Tortuosity of Blood VesselsCode0
A comparison of streaming models and data augmentation methods for robust speech recognition0
Semi-supervised transfer learning for language expansion of end-to-end speech recognition models to low-resource languages0
Self-Supervised Class Incremental Learning0
Rethinking Drone-Based Search and Rescue with Aerial Person DetectionCode1
FLSys: Toward an Open Ecosystem for Federated Learning Mobile Apps0
Guiding Generative Language Models for Data Augmentation in Few-Shot Text Classification0
TYolov5: A Temporal Yolov5 Detector Based on Quasi-Recurrent Neural Networks for Real-Time Handgun Detection in VideoCode1
Show:102550
← PrevPage 204 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified