TY - GEN
T1 - Improving humanoid locomotive performance with learnt approximated dynamics via Gaussian processes for regression
AU - Morimoto, Jun
AU - Atkeson, Christopher G.
AU - Endo, Gen
AU - Cheng, Gordon
PY - 2007
Y1 - 2007
N2 - We propose to improve the locomotive performance of humanoid robots by using approximated biped stepping and walking dynamics with reinforcement learning (RL). Although RL is a useful non-linear optimizer, it is usually difficult to apply RL to real robotic systems - due to the large number of iterations required to acquire suitable policies. In this study, we first approximated the dynamics by using data from a real robot, and then applied the estimated dynamics in RL in order to improve stepping and walking policies. Gaussian processes were used to approximate the dynamics. By using Gaussian processes, we could estimate a probability distribution of a target function with a given covariance function. Thus, RL can take the uncertainty of the approximated dynamics into account throughout the learning process. We show that we can improve stepping and walking policies by using a RL method with the approximated models both in simulated and real environments. Experimental validation on a real humanoid robot of the proposed learning approach is presented.
AB - We propose to improve the locomotive performance of humanoid robots by using approximated biped stepping and walking dynamics with reinforcement learning (RL). Although RL is a useful non-linear optimizer, it is usually difficult to apply RL to real robotic systems - due to the large number of iterations required to acquire suitable policies. In this study, we first approximated the dynamics by using data from a real robot, and then applied the estimated dynamics in RL in order to improve stepping and walking policies. Gaussian processes were used to approximate the dynamics. By using Gaussian processes, we could estimate a probability distribution of a target function with a given covariance function. Thus, RL can take the uncertainty of the approximated dynamics into account throughout the learning process. We show that we can improve stepping and walking policies by using a RL method with the approximated models both in simulated and real environments. Experimental validation on a real humanoid robot of the proposed learning approach is presented.
UR - http://www.scopus.com/inward/record.url?scp=51349148802&partnerID=8YFLogxK
U2 - 10.1109/IROS.2007.4399485
DO - 10.1109/IROS.2007.4399485
M3 - Conference contribution
AN - SCOPUS:51349148802
SN - 1424409128
SN - 9781424409129
T3 - IEEE International Conference on Intelligent Robots and Systems
SP - 4234
EP - 4240
BT - Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2007
T2 - 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2007
Y2 - 29 October 2007 through 2 November 2007
ER -