SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 451475 of 8378 papers

TitleStatusHype
TimeDistill: Efficient Long-Term Time Series Forecasting with MLP via Cross-Architecture Distillation0
External Large Foundation Model: How to Efficiently Serve Trillions of Parameters for Online Ads Recommendation0
CrossFuse: Learning Infrared and Visible Image Fusion by Cross-Sensor Top-K Vision Alignment and Beyond0
Reducing false positives in strong lens detection through effective augmentation and ensemble learning0
Humanoid-VLA: Towards Universal Humanoid Control with Visual Integration0
Image compositing is all you need for data augmentation0
AS-GCL: Asymmetric Spectral Augmentation on Graph Contrastive Learning0
Democratizing Large Language Model-Based Graph Data Augmentation via Latent Knowledge GraphsCode0
Integrating Arithmetic Learning Improves Mathematical Reasoning in Smaller Models0
MVCNet: Multi-View Contrastive Network for Motor Imagery ClassificationCode1
Fake It Till You Make It: Using Synthetic Data and Domain Knowledge for Improved Text-Based Learning for LGE Detection0
Generative AI Enabled Robust Data Augmentation for Wireless Sensing in ISAC Networks0
Myna: Masking-Based Contrastive Learning of Musical RepresentationsCode1
Diversity-Oriented Data Augmentation with Large Language Models0
Understanding In-Context Machine Translation for Low-Resource Languages: A Case Study on ManchuCode1
Building A Proof-Oriented Programmer That Is 64% Better Than GPT-4o Under Data Scarcity0
SpeechT: Findings of the First Mentorship in Speech Translation0
CARMA: Enhanced Compositionality in LLMs via Advanced Regularisation and Mutual Information Alignment0
ReLearn: Unlearning via Learning for Large Language ModelsCode1
AudioSpa: Spatializing Sound Events with Text0
Generating Skyline Datasets for Data Science Models0
A Mathematics Framework of Artificial Shifted Population Risk and Its Further Understanding Related to Consistency RegularizationCode0
NeuroAMP: A Novel End-to-end General Purpose Deep Neural Amplifier for Personalized Hearing Aids0
Rao-Blackwell Gradient Estimators for Equivariant Denoising Diffusion0
Causal Information Prioritization for Efficient Reinforcement Learning0
Show:102550
← PrevPage 19 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified