Non-Vacuous Generalisation Bounds for Shallow Neural Networks. Biggs, F. & Guedj, B. In Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., & Sabato, S., editors, Proceedings of the 39th International Conference on Machine Learning [ICML], volume 162, of Proceedings of Machine Learning Research, pages 1963–1981, July, 2022. PMLR.
Non-Vacuous Generalisation Bounds for Shallow Neural Networks [link]Paper  Non-Vacuous Generalisation Bounds for Shallow Neural Networks [link]Arxiv  Non-Vacuous Generalisation Bounds for Shallow Neural Networks [pdf]Pdf  Non-Vacuous Generalisation Bounds for Shallow Neural Networks [link]Code  Non-Vacuous Generalisation Bounds for Shallow Neural Networks [link]Video  Non-Vacuous Generalisation Bounds for Shallow Neural Networks [pdf]Slides  Non-Vacuous Generalisation Bounds for Shallow Neural Networks [png]Poster  Non-Vacuous Generalisation Bounds for Shallow Neural Networks [link]Slideslive  abstract   bibtex   1 download  
We focus on a specific class of shallow neural networks with a single hidden layer, namely those with $L_2$-normalised data and either a sigmoid-shaped Gaussian error function (“erf”) activation or a Gaussian Error Linear Unit (GELU) activation. For these networks, we derive new generalisation bounds through the PAC-Bayesian theory; unlike most existing such bounds they apply to neural networks with deterministic rather than randomised parameters. Our bounds are empirically non-vacuous when the network is trained with vanilla stochastic gradient descent on MNIST and Fashion-MNIST.

Downloads: 1