TY - JOUR
T1 - DEEP LEARNING APPROXIMATION OF DIFFEOMORPHISMS VIA LINEAR-CONTROL SYSTEMS
AU - Scagliotti, Alessandro
N1 - Publisher Copyright:
© 2023, American Institute of Mathematical Sciences. All rights reserved.
PY - 2023/9
Y1 - 2023/9
N2 - In this paper we propose a Deep Learning architecture to approximate diffeomorphisms diffeotopic to the identity. We consider a control system of the form ẋ =∑l i=1Fi(x)ui, with linear dependence in the controls, and we use the corresponding flow to approximate the action of a diffeomorphism on a compact ensemble of points. Despite the simplicity of the control system, it has been recently shown that a Universal Approximation Property holds. The problem of minimizing the sum of the training error and of a regularizing term induces a gradient flow in the space of admissible controls. A possible training procedure for the discrete-time neural network consists in projecting the gradient flow onto a finite-dimensional subspace of the admissible controls. An alternative approach relies on an iterative method based on Pontryagin Maximum Principle for the numerical resolution of Optimal Control problems. Here the maximization of the Hamiltonian can be carried out with an extremely low computational effort, owing to the linear dependence of the system in the control variables. Finally, we use tools from Γ-convergence to provide an estimate of the expected generalization error.
AB - In this paper we propose a Deep Learning architecture to approximate diffeomorphisms diffeotopic to the identity. We consider a control system of the form ẋ =∑l i=1Fi(x)ui, with linear dependence in the controls, and we use the corresponding flow to approximate the action of a diffeomorphism on a compact ensemble of points. Despite the simplicity of the control system, it has been recently shown that a Universal Approximation Property holds. The problem of minimizing the sum of the training error and of a regularizing term induces a gradient flow in the space of admissible controls. A possible training procedure for the discrete-time neural network consists in projecting the gradient flow onto a finite-dimensional subspace of the admissible controls. An alternative approach relies on an iterative method based on Pontryagin Maximum Principle for the numerical resolution of Optimal Control problems. Here the maximization of the Hamiltonian can be carried out with an extremely low computational effort, owing to the linear dependence of the system in the control variables. Finally, we use tools from Γ-convergence to provide an estimate of the expected generalization error.
KW - Deep Learning
KW - Pontryagin Maximum Principle
KW - gradient flow
KW - linear-control system
KW - Γ-convergence
UR - http://www.scopus.com/inward/record.url?scp=85147909242&partnerID=8YFLogxK
U2 - 10.3934/mcrf.2022036
DO - 10.3934/mcrf.2022036
M3 - Article
AN - SCOPUS:85147909242
SN - 2156-8472
VL - 13
SP - 1226
EP - 1257
JO - Mathematical Control and Related Fields
JF - Mathematical Control and Related Fields
IS - 3
ER -