SOTAVerified

Second-order methods

Use second-order statistics to process data.

Papers

Showing 150 of 181 papers

TitleStatusHype
Towards Practical Second-Order Optimizers in Deep Learning: Insights from Fisher Information AnalysisCode2
Automatic Gradient Descent: Deep Learning without HyperparametersCode2
Differentiable Expected Hypervolume Improvement for Parallel Multi-Objective Bayesian OptimizationCode2
SASSHA: Sharpness-aware Adaptive Second-order Optimization with Stable Hessian ApproximationCode1
Revisiting Scalable Hessian Diagonal Approximations for Applications in Reinforcement LearningCode1
Can We Remove the Square-Root in Adaptive Gradient Methods? A Second-Order PerspectiveCode1
Structured Inverse-Free Natural Gradient: Memory-Efficient & Numerically-Stable KFACCode1
MKOR: Momentum-Enabled Kronecker-Factor-Based Optimizer Using Rank-1 UpdatesCode1
Symmetry Teleportation for Accelerated OptimizationCode1
Near out-of-distribution detection for low-resolution radar micro-Doppler signaturesCode1
Communication-Efficient Stochastic Zeroth-Order Optimization for Federated LearningCode1
Second-Order Neural ODE OptimizerCode1
M-FAC: Efficient Matrix-Free Approximations of Second-Order InformationCode1
ADAHESSIAN: An Adaptive Second Order Optimizer for Machine LearningCode1
MiLeNAS: Efficient Neural Architecture Search via Mixed-Level ReformulationCode1
Low Rank Saddle Free Newton: A Scalable Method for Stochastic Nonconvex OptimizationCode1
Second-Order Stochastic Optimization for Machine Learning in Linear TimeCode1
NysAct: A Scalable Preconditioned Gradient Descent using Nystrom ApproximationCode0
KOALA++: Efficient Kalman-Based Optimization of Neural Networks with Gradient-Covariance Products0
Accelerated Training of Federated Learning via Second-Order Methods0
Representation Meets Optimization: Training PINNs and PIKANs for Gray-Box Discovery in Systems Pharmacology0
FedOSAA: Improving Federated Learning with One-Step Anderson Acceleration0
Online Covariance Matrix Estimation in Sketched Newton MethodsCode0
A Survey of Optimization Methods for Training DL Models: Theoretical Perspective on Convergence and Generalization0
Distributed Quasi-Newton Method for Fair and Fast Federated Learning0
Preconditioners for the Stochastic Training of Neural Fields0
GP-FL: Model-Based Hessian Estimation for Second-Order Over-the-Air Federated Learning0
A Learn-to-Optimize Approach for Coordinate-Wise Step Sizes for Quasi-Newton Methods0
Gradient Norm Regularization Second-Order Algorithms for Solving Nonconvex-Strongly Concave Minimax Problems0
Don't Be So Positive: Negative Step Sizes in Second-Order Methods0
Exact, Tractable Gauss-Newton Optimization in Deep Reversible Architectures Reveal Poor GeneralizationCode0
Improving Stochastic Cubic Newton with Momentum0
Second-Order Min-Max Optimization with Lazy Hessians0
Unlocking FedNL: Self-Contained Compute-Optimized Implementation0
Adversarial Vulnerability as a Consequence of On-Manifold Inseparibility0
FLeNS: Federated Learning with Enhanced Nesterov-Newton SketchCode0
Alternating Iteratively Reweighted _1 and Subspace Newton Algorithms for Nonconvex Sparse OptimizationCode0
Fed-Sophia: A Communication-Efficient Second-Order Federated Learning Algorithm0
Adaptive and Optimal Second-order Optimistic Methods for Minimax Optimization0
Stochastic Newton Proximal Extragradient Method0
Kronecker-Factored Approximate Curvature for Physics-Informed Neural Networks0
A Gauss-Newton Approach for Min-Max Optimization in Generative Adversarial NetworksCode0
Inverse-Free Fast Natural Gradient Descent Method for Deep Learning0
SGD with Partial Hessian for Deep Neural Networks OptimizationCode0
Second Order Methods for Bandit Optimization and Control0
The Challenges of the Nonlinear Regime for Physics-Informed Neural Networks0
On The Temporal Domain of Differential Equation Inspired Graph Neural Networks0
Krylov Cubic Regularized Newton: A Subspace Second-Order Method with Dimension-Free Convergence Rate0
Adapting Newton's Method to Neural Networks through a Summary of Higher-Order DerivativesCode0
A Computationally Efficient Sparsified Online Newton MethodCode0
Show:102550
← PrevPage 1 of 4Next →

No leaderboard results yet.