SOTAVerified

Neural Architecture Search

Neural architecture search (NAS) is a technique for automating the design of artificial neural networks (ANN), a widely used model in the field of machine learning. NAS essentially takes the process of a human manually tweaking a neural network and learning what works well, and automates this task to discover more complex architectures.

Image Credit : NAS with Reinforcement Learning

Papers

Showing 901950 of 1915 papers

TitleStatusHype
BINAS: Bilinear Interpretable Neural Architecture SearchCode0
Towards a Robust Differentiable Architecture Search under Label Noise0
Grafting Transformer on Automatically Designed Convolutional Neural Network for Hyperspectral Image ClassificationCode1
ProxyBO: Accelerating Neural Architecture Search via Bayesian Optimization with Zero-cost Proxies0
DARTS for Inverse Problems: a Study on Stability0
NAS-HPO-Bench-II: A Benchmark Dataset on Joint Optimization of Convolutional Neural Network Architecture and Training HyperparametersCode1
Growing Representation Learning0
NeuralArTS: Structuring Neural Architecture Search with Type Theory0
GradSign: Model Performance Inference with Theoretical InsightsCode0
DPNAS: Neural Architecture Search for Deep Learning with Differential PrivacyCode0
CONetV2: Efficient Auto-Channel Size Optimization for CNNsCode0
Improving the sample-efficiency of neural architecture search with reinforcement learningCode0
NAS-Bench-360: Benchmarking Neural Architecture Search on Diverse TasksCode1
On the Security Risks of AutoMLCode0
Across-Task Neural Architecture Search via Meta Learning0
Neural Architecture Search for Efficient Uncalibrated Deep Photometric Stereo0
ZARTS: On Zero-order Optimization for Neural Architecture Search0
SuperShaper: Task-Agnostic Super Pre-training of BERT Models with Variable Hidden Dimensions0
A Genetic Programming Approach To Zero-Shot Neural Architecture Ranking0
Accelerating Multi-Objective Neural Architecture Search by Random-Weight Evaluation0
Dataset Condensation with Distribution MatchingCode1
Conceptual Expansion Neural Architecture Search (CENAS)0
Multi-objective Optimization by Learning Space PartitionsCode1
Max and Coincidence Neurons in Neural Networks0
An Analysis of Super-Net Heuristics in Weight-Sharing NAS0
Inexact bilevel stochastic gradient methods for constrained and unconstrained lower-level problemsCode0
DAAS: Differentiable Architecture and Augmentation Policy Search0
What to expect of hardware metric predictors in NAS0
Ranking Convolutional Architectures by their Feature Extraction Capabilities0
Personalized Neural Architecture Search for Federated Learning0
Tr-NAS: Memory-Efficient Neural Architecture Search with Transferred Blocks0
NASPY: Automated Extraction of Automated Machine Learning Models0
NAS-Bench-Zero: A Large Scale Dataset for Understanding Zero-Shot Neural Architecture Search0
Multi-objective optimization for Hardware-aware Neural Architecture Search0
A Novel Watermarking Framework for Ownership Verification of DNN Architectures0
Learning Efficient Image Super-Resolution Networks via Structure-Regularized Pruning0
ZenDet: Revisiting Efficient Object Detection Backbones from Zero-Shot Neural Architecture Search0
ConCoDE: Hard-constrained Differentiable Co-Exploration Method for Neural Architectures and Hardware Accelerators0
FedNAS: Federated Deep Learning via Neural Architecture Search0
A Transferable General-Purpose Predictor for Neural Architecture Search0
Multi-objective Optimization by Learning Space Partition0
Neural Architecture Search via Ensemble-based Knowledge Distillation0
Hardware-Aware Network Transformation0
Automated Mobile Attention KPConv Networks via A Wide & Deep Predictor0
Picking up the pieces: separately evaluating supernet training and architecture selection0
AutoCoG: A Unified Data-Modal Co-Search Framework for Graph Neural Networks0
SUMNAS: Supernet with Unbiased Meta-Features for Neural Architecture Search0
NASViT: Neural Architecture Search for Efficient Vision Transformers with Gradient Conflict aware Supernet TrainingCode1
L^2NAS: Learning to Optimize Neural Architectures via Continuous-Action Reinforcement Learning0
Profiling Neural Blocks and Design Spaces for Mobile Neural Architecture SearchCode0
Show:102550
← PrevPage 19 of 39Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1SPOS (ProxylessNAS (GPU) latency)Accuracy75.3Unverified
2SPOS (FBNet-C latency)Accuracy75.1Unverified
3SPOS (block search + channel search)Accuracy74.7Unverified
4MUXNet-xsTop-1 Error Rate33.3Unverified
5FBNetV2-F1Top-1 Error Rate31.7Unverified
6LayerNAS-60MTop-1 Error Rate31Unverified
7NASGEPTop-1 Error Rate29.51Unverified
8MUXNet-sTop-1 Error Rate28.4Unverified
9NN-MASS-ATop-1 Error Rate27.1Unverified
10FBNetV2-F3Top-1 Error Rate26.8Unverified
#ModelMetricClaimedVerifiedStatus
1CR-LSOAccuracy (Test)46.98Unverified
2Shapley-NASAccuracy (Test)46.85Unverified
3β-RDARTS-L2Accuracy (Test)46.71Unverified
4β-SDARTS-RSAccuracy (Test)46.71Unverified
5ASE-NAS+Accuracy (Val)46.66Unverified
6NARAccuracy (Test)46.66Unverified
7BaLeNAS-TFAccuracy (Test)46.54Unverified
8AG-NetAccuracy (Test)46.42Unverified
9Local searchAccuracy (Test)46.38Unverified
10NASBOTAccuracy (Test)46.37Unverified
#ModelMetricClaimedVerifiedStatus
1Balanced MixtureAccuracy (% )91.55Unverified
2GDASTop-1 Error Rate3.4Unverified
3Bonsai-NetTop-1 Error Rate3.35Unverified
4Net2 (2)Top-1 Error Rate3.3Unverified
5μDARTSTop-1 Error Rate3.28Unverified
6NN-MASS- CIFAR-CTop-1 Error Rate3.18Unverified
7NN-MASS- CIFAR-ATop-1 Error Rate3Unverified
8DARTS (first order)Top-1 Error Rate3Unverified
9NASGEPTop-1 Error Rate2.82Unverified
10AlphaX-1 (cutout NASNet)Top-1 Error Rate2.82Unverified