Initial state training procedure improves dynamic recurrent networks with time-dependent weights

Lutz Leistritz, Miroslaw Galicki, Herbert Witte, Eberhard Kochs

Research output: Contribution to journalArticlepeer-review

4 Scopus citations

Abstract

The problem of learning multiple continuous trajectories by means of recurrent neural networks with (in general) time-varying weights is addressed in this study. The learning process is transformed into an optimal control framework where both the weights and the initial network state to be found are treated as controls. For such a task, a new learning algorithm is proposed which is based on a variational formulation of Pontryagin's maximum principle. The convergence of this algorithm, under reasonable assumptions, is also investigated. Numerical examples of learning nontrivial two-class problems are presented which demonstrate the efficiency of the approach proposed.

Original languageEnglish
Pages (from-to)1513-1518
Number of pages6
JournalIEEE Transactions on Neural Networks
Volume12
Issue number6
DOIs
StatePublished - Nov 2001

Keywords

  • Dynamic neural networks
  • Initial network states
  • Optimal control
  • Trajectory learning

Fingerprint

Dive into the research topics of 'Initial state training procedure improves dynamic recurrent networks with time-dependent weights'. Together they form a unique fingerprint.

Cite this