TY - JOUR
T1 - Localized active learning of Gaussian process state space models
AU - Capone, Alexandre
AU - Noske, Gerrit
AU - Umlauft, Jonas
AU - Beckers, Thomas
AU - Lederer, Armin
AU - Hirche, Sandra
N1 - Publisher Copyright:
© 2020 A. Capone, G. Noske, J. Umlauft, T. Beckers, A. Lederer & S. Hirche.
PY - 2020
Y1 - 2020
N2 - While most dynamic system exploration techniques aim to achieve a globally accurate model, this is generally unsuited for systems with unbounded state spaces. Furthermore, many applications do not require a globally accurate model, e.g., local stabilization tasks. In this paper, we propose an active learning strategy for Gaussian process state space models that aims to obtain an accurate model on a bounded subset of the state-action space. Our approach aims to maximize the mutual information of the exploration trajectories with respect to a discretization of the region of interest. By employing model predictive control, the proposed technique integrates information collected during exploration and adaptively improves its exploration strategy. To enable computational tractability, we decouple the choice of most informative data points from the model predictive control optimization step. This yields two optimization problems that can be solved in parallel. We apply the proposed method to explore the state space of various dynamical systems and compare our approach to a commonly used entropy-based exploration strategy. In all experiments, our method yields a better model within the region of interest than the entropy-based method.
AB - While most dynamic system exploration techniques aim to achieve a globally accurate model, this is generally unsuited for systems with unbounded state spaces. Furthermore, many applications do not require a globally accurate model, e.g., local stabilization tasks. In this paper, we propose an active learning strategy for Gaussian process state space models that aims to obtain an accurate model on a bounded subset of the state-action space. Our approach aims to maximize the mutual information of the exploration trajectories with respect to a discretization of the region of interest. By employing model predictive control, the proposed technique integrates information collected during exploration and adaptively improves its exploration strategy. To enable computational tractability, we decouple the choice of most informative data points from the model predictive control optimization step. This yields two optimization problems that can be solved in parallel. We apply the proposed method to explore the state space of various dynamical systems and compare our approach to a commonly used entropy-based exploration strategy. In all experiments, our method yields a better model within the region of interest than the entropy-based method.
KW - Bayesian inference
KW - data-driven control
KW - exploration
KW - model predictive control
UR - http://www.scopus.com/inward/record.url?scp=85090081659&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85090081659
SN - 2640-3498
VL - 120
SP - 490
EP - 499
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
T2 - 2nd Annual Conference on Learning for Dynamics and Control, L4DC 2020
Y2 - 10 June 2020 through 11 June 2020
ER -