Journals
Resources
About Us
Open Access

Approximation Results for Gradient Flow Trained Neural Networks

Year:    2024

Author:    Gerrit Welper

Journal of Machine Learning, Vol. 3 (2024), Iss. 2 : pp. 107–175

Abstract

The paper contains approximation guarantees for neural networks that are trained with gradient flow, with error measured in the continuous $L_2(\mathbb{S}^{d−1 )}$-norm on the $d$-dimensional unit sphere and targets that are Sobolev smooth. The networks are fully connected of constant depth and increasing width. We show gradient flow convergence based on a neural tangent kernel (NTK) argument for the non-convex optimization of the second but last layer. Unlike standard NTK analysis, the continuous error norm implies an under-parametrized regime, possible by the natural smoothness assumption required for approximation. The typical over-parametrization re-enters the results in form of a loss in approximation rate relative to established approximation methods for Sobolev smooth functions.

Journal Article Details

Publisher Name:    Global Science Press

Language:    English

DOI:    https://doi.org/10.4208/jml.230924

Journal of Machine Learning, Vol. 3 (2024), Iss. 2 : pp. 107–175

Published online:    2024-01

AMS Subject Headings:   

Copyright:    COPYRIGHT: © Global Science Press

Pages:    69

Keywords:    Deep neural networks Approximation Gradient descent Neural tangent kernel.

Author Details

Gerrit Welper