SOTAVerified

Second-order methods

Use second-order statistics to process data.

Papers

Showing 101150 of 181 papers

TitleStatusHype
FedNL: Making Newton-Type Methods Applicable to Federated Learning0
Tensor Normal Training for Deep Learning ModelsCode0
Exact Stochastic Second Order Deep Learning0
Quasi-Newton Quasi-Monte Carlo for variational Bayes0
Research of Damped Newton Stochastic Gradient Descent Method for Neural Network Training0
A Distributed Optimisation Framework Combining Natural Gradient with Hessian-Free for Discriminative Sequence Training0
Learning-Augmented Sketches for Hessians0
Distributed Second Order Methods with Fast Rates and Compressed Communication0
Kronecker-factored Quasi-Newton Methods for Deep Learning0
A Chaos Theory Approach to Understand Neural Network Optimization0
Adaptive Single-Pass Stochastic Gradient Descent in Input Sparsity Time0
Noise and Fluctuation of Finite Learning Rate Stochastic Gradient Descent0
Utility Maximization for Large-Scale Cell-Free Massive MIMO Downlink0
Second-order Neural Network Training Using Complex-step Directional Derivative0
Debiasing Distributed Second Order Optimization with Surrogate Sketching and Scaled Regularization0
Second-Order Information in Non-Convex Stochastic Optimization: Power and Limitations0
When Does Preconditioning Help or Hurt Generalization?0
A block coordinate descent optimizer for classification problems exploiting convexity0
Enhance Curvature Information by Structured Stochastic Quasi-Newton Methods0
Learning Rates as a Function of Batch Size: A Random Matrix Theory Approach to Neural Network TrainingCode0
SONIA: A Symmetric Blockwise Truncated Optimization Algorithm0
Asymptotic Analysis of Conditioned Stochastic Gradient DescentCode0
On the Promise of the Stochastic Generalized Gauss-Newton Method for Training DNNsCode0
Critical Point-Finding Methods Reveal Gradient-Flat Regions of Deep Network Losses0
Stochastic Subspace Cubic Newton Method0
DDPNOpt: Differential Dynamic Programming Neural Optimizer0
Newtonian Monte Carlo: single-site MCMC meets second-order gradient methodsCode0
Curvature-corrected learning dynamics in deep neural networks0
Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic RatesCode0
Hierarchical model-based policy optimization: from actions to action sequences and back0
Implementation of a modified Nesterov's Accelerated quasi-Newton Method on Tensorflow0
Fast and Furious Convergence: Stochastic Second Order Methods under InterpolationCode0
EXACT ANALYSIS OF CURVATURE CORRECTED LEARNING DYNAMICS IN DEEP LINEAR NETWORKS0
Quasi-Newton Optimization Methods For Deep Learning Applications0
Practical Newton-Type Distributed Learning using Gradient Based Approximations0
Meta-descent for Online, Continual Prediction0
SGD momentum optimizer with step estimation by online parabola modelCode0
Limitations of the Empirical Fisher Approximation for Natural Gradient DescentCode0
Gram-Gauss-Newton Method: Learning Overparameterized Neural Networks for Regression Problems0
LIBS2ML: A Library for Scalable Second Order Machine Learning AlgorithmsCode0
Improving SGD convergence by online linear regression of gradients in multiple statistically relevant directionsCode0
Stochastic Trust Region Inexact Newton Method for Large-scale Machine LearningCode0
Bilinear Parameterization For Differentiable Rank-Regularization0
Deep Reinforcement Learning via L-BFGS Optimization0
Large batch size training of neural networks with adversarial training and second-order informationCode0
Stochastic Second-order Methods for Non-convex Optimization with Inexact Hessian and Gradient0
A Distributed Second-Order Algorithm You Can Trust0
Online Second Order Methods for Non-Convex Stochastic OptimizationsCode0
GPU Accelerated Sub-Sampled Newton's Method0
The Many Faces of Exponential Weights in Online Learning0
Show:102550
← PrevPage 3 of 4Next →

No leaderboard results yet.