SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 62766300 of 8378 papers

TitleStatusHype
Long-tailed Recognition by Learning from Latent Categories0
Long-Term Modeling of Financial Machine Learning for Active Portfolio Management0
Long Term Object Detection and Tracking in Collaborative Learning Environments0
Long-time predictive modeling of nonlinear dynamical systems using neural networks0
Lorentz Local Canonicalization: How to Make Any Network Lorentz-Equivariant0
LOST: A Mental Health Dataset of Low Self-esteem in Reddit Posts0
Lost in Translation: Modern Neural Networks Still Struggle With Small Realistic Image Transformations0
Lost in Transmission: On the Impact of Networking Corruptions on Video Machine Learning Models0
Low-Complexity Acoustic Scene Classification Using Data Augmentation and Lightweight ResNet0
Low-complexity deep learning frameworks for acoustic scene classification0
Low-Complexity Own Voice Reconstruction for Hearables with an In-Ear Microphone0
Low-cost and high-performance data augmentation for deep-learning-based skin lesion classification0
Low-data? No problem: low-resource, language-agnostic conversational text-to-speech via F0-conditioned data augmentation0
Low-rank representation of head impact kinematics: A data-driven emulator0
Low-Resolution Neural Networks0
Low-resource expressive text-to-speech using data augmentation0
Low Resource German ASR with Untranscribed Data Spoken by Non-native Children -- INTERSPEECH 2021 Shared Task SPAPL System0
Low resource language dataset creation, curation and classification: Setswana and Sepedi -- Extended Abstract0
Zero-shot and Few-shot Learning with Knowledge Graphs: A Comprehensive Survey0
Pretraining by Backtranslation for End-to-end ASR in Low-Resource Settings0
Low-Resource Text-to-Speech Using Specific Data and Noise Augmentation0
Low-Resource Vision Challenges for Foundation Models0
Low-shot Learning via Covariance-Preserving Adversarial Augmentation Networks0
LPC Augment: An LPC-Based ASR Data Augmentation Algorithm for Low and Zero-Resource Children's Dialects0
LS-Net: Fast Single-Shot Line-Segment Detector0
Show:102550
← PrevPage 252 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified