SOTAVerified

Data Augmentation

Data augmentation involves techniques used for increasing the amount of data, based on different modifications, to expand the amount of examples in the original dataset. Data augmentation not only helps to grow the dataset but it also increases the diversity of the dataset. When training machine learning models, data augmentation acts as a regularizer and helps to avoid overfitting.

Data augmentation techniques have been found useful in domains like NLP and computer vision. In computer vision, transformations like cropping, flipping, and rotation are used. In NLP, data augmentation techniques can include swapping, deletion, random insertion, among others.

Further readings:

( Image credit: Albumentations )

Papers

Showing 51765200 of 8378 papers

TitleStatusHype
Long Term Stock Prediction based on Financial StatementsCode0
Tencent Translation System for the WMT21 News Translation Task0
The LMU Munich System for the WMT 2021 Large-Scale Multilingual Machine Translation Shared Task0
Contrastive Learning for Context-aware Neural Machine Translation Using Coreference Information0
HW-TSC’s Participation at WMT 2021 Quality Estimation Shared Task0
TermMind: Alibaba’s WMT21 Machine Translation Using Terminologies Task Submission0
NVIDIA NeMo’s Neural Machine Translation Systems for English-German and English-Russian News and Biomedical Tasks at WMT210
SpanAlign: Efficient Sequence Tagging Annotation Projection into Translated Data applied to Cross-Lingual Opinion Mining0
DMix: Distance Constrained Interpolative Mixup0
Sequence Mixup for Zero-Shot Cross-Lingual Part-Of-Speech Tagging0
HypMix: Hyperbolic Interpolative Data AugmentationCode1
Learning Data Augmentation Schedules for Natural Language ProcessingCode0
A Semantic Feature-Wise Transformation Relation Network for Automatic Short Answer Grading0
Reinforced Counterfactual Data Augmentation for Dual Sentiment ClassificationCode0
AESOP: Paraphrase Generation with Adaptive Syntactic ControlCode1
Simple Conversational Data Augmentation for Semi-supervised Abstractive Dialogue SummarizationCode1
Machine Reading Comprehension as Data Augmentation: A Case Study on Implicit Event Argument Extraction0
Gradient-Based Adversarial Factual Consistency Evaluation for Abstractive Summarization0
Semantics-Preserved Data Augmentation for Aspect-Based Sentiment Analysis0
TADPOLE: Task ADapted Pre-Training via AnOmaLy DEtection0
Can Question Generation Debias Question Answering Models? A Case Study on Question–Context Lexical Overlap0
PDALN: Progressive Domain Adaptation over a Pre-trained Model for Low-Resource Cross-Domain Named Entity Recognition0
Can We Improve Model Robustness through Secondary Attribute Counterfactuals?0
Unsupervised Paraphrasing Consistency Training for Low Resource Named Entity Recognition0
NDH-Full: Learning and Evaluating Navigational Agents on Full-Length DialogueCode0
Show:102550
← PrevPage 208 of 336Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeiT-B (+MixPro)Accuracy (%)82.9Unverified
2ResNet-200 (DeepAA)Accuracy (%)81.32Unverified
3DeiT-S (+MixPro)Accuracy (%)81.3Unverified
4ResNet-200 (Fast AA)Accuracy (%)80.6Unverified
5ResNet-200 (UA)Accuracy (%)80.4Unverified
6ResNet-200 (AA)Accuracy (%)80Unverified
7ResNet-50 (DeepAA)Accuracy (%)78.3Unverified
8ResNet-50 (TA wide)Accuracy (%)78.07Unverified
9ResNet-50 (LoRot-E)Accuracy (%)77.72Unverified
10ResNet-50 (LoRot-I)Accuracy (%)77.71Unverified
#ModelMetricClaimedVerifiedStatus
1WideResNet-40-2 (Faster AA)Percentage error3.7Unverified
2Shake-Shake (26 2×32d) (Faster AA)Percentage error2.7Unverified
3WideResNet-28-10 (Faster AA)Percentage error2.6Unverified
4Shake-Shake (26 2×112d) (Faster AA)Percentage error2Unverified
5Shake-Shake (26 2×96d) (Faster AA)Percentage error2Unverified
#ModelMetricClaimedVerifiedStatus
1DiffAugClassification Accuracy92.7Unverified
2PaCMAPClassification Accuracy85.3Unverified
3hNNEClassification Accuracy77.4Unverified
4TopoAEClassification Accuracy74.6Unverified