Spectral Neural Networks: Approximation Theory and Optimization Landscape

Author(s)

,
&

Abstract

There is a large variety of machine learning methodologies that are based on the extraction of spectral geometric information from data. However, the implementations of many of these methods often depend on traditional eigensolvers, which present limitations when applied in practical online big data scenarios. To address some of these challenges, researchers have proposed different strategies for training neural networks as alternatives to traditional eigensolvers, with one such approach known as spectral neural network (SNN). In this paper, we investigate key theoretical aspects of SNN. First, we present quantitative insights into the tradeoff between the number of neurons and the amount of spectral geometric information a neural network learns. Second, we initiate a theoretical exploration of the optimization landscape of SNN’s objective to shed light on the training dynamics of SNN. Unlike typical studies of convergence to global solutions of NN training dynamics, SNN presents an additional complexity due to its non-convex ambient loss function, a feature that is common in unsupervised learning settings.

Author Biographies

  • Chenghui Li

    Department of Statistics, University of Wisconsin Madison, Madison, Wisconsin 53706, USA

  • Rishi Sonthalia

    Department of Mathematics, Boston College, Boston, MA 02109, USA

  • Nicolás García Trillos

    Department of Statistics, University of Wisconsin Madison, Madison, Wisconsin 53706, USA

About this article

Abstract View

  • 123

Pdf View

  • 206

DOI

10.4208/jml.240322