Neural Networks as Inter-Domain Inducing Points
2020-11-23pproximateinference AABI Symposium 2021Unverified0· sign in to hype
Shengyang Sun, Jiaxin Shi, Roger Baker Grosse
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Equivalences between infinite neural networks and Gaussian processes have been established for explaining the functional prior and training dynamics of deep learning models. In this paper we cast the hidden units of finite-width neural networks as the inter-domain inducing points of a kernel, then a one-hidden-layer network becomes a kernel regression model. For dot-product kernels on both R^d and S^d−1, we derive the kernel functions for inducing points. Empirically we conduct toy experiments to validate the proposed approaches.