Efficient identification of wide shallow neural networks with biases

Massimo Fornasier, Timo Klock, Marco Mondelli, Michael Rauchensteiner

Research output: Contribution to journalArticlepeer-review

Abstract

The identification of the parameters of a neural network from finite samples of input-output pairs is often referred to as the teacher-student model, and this model has represented a popular framework for understanding training and generalization. Even if the problem is NP-complete in the worst case, a rapidly growing literature – after adding suitable distributional assumptions – has established finite sample identification of two-layer networks with a number of neurons m=O(D), D being the input dimension. For the range D<m<D2 the problem becomes harder, and truly little is known for networks parametrized by biases as well. This paper fills the gap by providing efficient algorithms and rigorous theoretical guarantees of finite sample identification for such wider shallow networks with biases. Our approach is based on a two-step pipeline: first, we recover the direction of the weights, by exploiting second order information; next, we identify the signs by suitable algebraic evaluations, and we recover the biases by empirical risk minimization via gradient descent. Numerical results demonstrate the effectiveness of our approach.

Original languageEnglish
Article number101749
JournalApplied and Computational Harmonic Analysis
Volume77
DOIs
StatePublished - Jun 2025
Externally publishedYes

Fingerprint

Dive into the research topics of 'Efficient identification of wide shallow neural networks with biases'. Together they form a unique fingerprint.

Cite this