Non-asymptotic approximations of Gaussian neural networks via second-order Poincaré inequalities

Abstract

There is a growing interest on large-width asymptotic properties of Gaussian neural networks (NNs), namely NNs whose weights are initialized according to Gaussian distributions. A well-established result is that, as the width goes to infinity, a Gaussian NN converges in distribution to a Gaussian stochastic process, which provides an asymptotic or qualitative Gaussian approximation of the NN. In this paper, we introduce some non-asymptotic or quantitative Gaussian approximations of Gaussian NNs, quantifying the approximation error with respect to some popular distances for (probability) distributions, e.g. the 1-Wasserstein distance, the total variation distance and the Kolmogorov-Smirnov distance. Our results rely on the use of second-order Gaussian Poincaré inequalities, which provide tight estimates of the approximation error, with optimal rates. This is a novel application of second-order Gaussian Poincaré inequalities, which are well-known in the probabilistic literature for being a powerful tool to obtain Gaussian approximations of general functionals of Gaussian stochastic processes. A generalization of our results to deep Gaussian NNs is discussed.

Publication
In the Proceedings of Machine Learning Research (Contributed talk at the 6th Symposium on Advances in Approximate Bayesian Inference)
Click the Cite button above to demo the feature to enable visitors to import publication metadata into their reference management software.
Alberto Bordino
Alberto Bordino
PhD Student in Statistics, Warwick CDT in Statistics

Third-year PhD candidate developing nonparametric, minimax-optimal methods for learning with missing or heterogeneous data.