SOTAVerified

Active Learning

Active Learning is a paradigm in supervised machine learning which uses fewer training examples to achieve better optimization by iteratively training a predictor, and using the predictor in each iteration to choose the training examples which will increase its chances of finding better configurations and at the same time improving the accuracy of the prediction model

Source: Polystore++: Accelerated Polystore System for Heterogeneous Workloads

Papers

Showing 201225 of 3073 papers

TitleStatusHype
ORIS: Online Active Learning Using Reinforcement Learning-based Inclusive Sampling for Robust Streaming Analytics System0
Active partitioning: inverting the paradigm of active learning0
Spectral-Spatial Transformer with Active Transfer Learning for Hyperspectral Image ClassificationCode1
Multi-Label Bayesian Active Learning with Inter-Label RelationshipsCode0
Maximally Separated Active Learning0
A Review of Bayesian Uncertainty Quantification in Deep Probabilistic Image Segmentation0
Integrating Deep Metric Learning with Coreset for Active Learning in 3D SegmentationCode0
Benchmarking Active Learning for NILM0
Influence functions and regularity tangents for efficient active learning0
Active Learning-Based Optimization of Hydroelectric Turbine Startup to Minimize Fatigue Damage0
LPLgrad: Optimizing Active Learning Through Gradient Norm Sample Selection and Auxiliary Model TrainingCode0
Integration of Active Learning and MCMC Sampling for Efficient Bayesian Calibration of Mechanical Properties0
Stream-Based Active Learning for Process Monitoring0
Active learning for efficient discovery of optimal gene combinations in the combinatorial perturbation spaceCode0
Progressive Generalization Risk Reduction for Data-Efficient Causal Effect EstimationCode0
MolParser: End-to-end Visual Recognition of Molecule Structures in the Wild0
Targeting Negative Flips in Active Learning using Validation SetsCode0
Learning Quantitative Automata Modulo Theories0
Deep Active Learning in the Open World0
GCI-ViTAL: Gradual Confidence Improvement with Vision Transformers for Active Learning on Label Noise0
FisherMask: Enhancing Neural Network Labeling Efficiency in Image Classification Using Fisher InformationCode0
Performance-Guided LLM Knowledge Distillation for Efficient Text Classification at Scale0
Hands-On Tutorial: Labeling with LLM and Human-in-the-Loop0
Constrained Multi-objective Bayesian Optimization through Optimistic Constraints EstimationCode0
An information-matching approach to optimal experimental design and active learning0
Show:102550
← PrevPage 9 of 123Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1TypiClustAccuracy93.2Unverified
2PT4ALAccuracy93.1Unverified
3Learning lossAccuracy91.01Unverified
4CoreGCNAccuracy90.7Unverified
5Core-setAccuracy89.92Unverified
6Random Baseline (Resnet18)Accuracy88.45Unverified
7Random Baseline (VGG16)Accuracy85.09Unverified