Reinforcement Learning for Operational Space Control
- 1 April 2007
- conference paper
- Published by Institute of Electrical and Electronics Engineers (IEEE)
- No. 10504729,p. 2111-2116
- https://doi.org/10.1109/robot.2007.363633
Abstract
While operational space control is of essential importance for robotics and well-understood from an analytical point of view, it can be prohibitively hard to achieve accurate control in face of modeling errors, which are inevitable in complex robots, e.g., humanoid robots. In such cases, learning control methods can offer an interesting alternative to analytical control algorithms. However, the resulting supervised learning problem is ill-defined as it requires to learn an inverse mapping of a usually redundant system, which is well known to suffer from the property of non-convexity of the solution space, i.e., the learning system could generate motor commands that try to steer the robot into physically impossible configurations. The important insight that many operational space control algorithms can be reformulated as optimal control problems, however, allows addressing this inverse learning problem in the framework of reinforcement learning. However, few of the known optimization or reinforcement learning algorithms can be used in online learning control for robots, as they are either prohibitively slow, do not scale to interesting domains of complex robots, or require trying out policies generated by random search, which are infeasible for a physical system. Using a generalization of the EM-based reinforcement learning framework suggested by Dayan and Hinton, we reduce the problem of learning with immediate rewards to a reward-weighted regression problem with an adaptive, integrated reward transformation for faster convergence. The resulting algorithm is efficient, learns smoothly without dangerous jumps in solution space, and works well in applications of complex high degree-of-freedom robots.Keywords
This publication has 13 references indexed in Scilit:
- Internal Models for Motor ControlPublished by Wiley ,2007
- Comparative experiments on task space control with redundancy resolutionPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2005
- A unifying methodology for the control of robotic systemsPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2005
- Introduction to Stochastic Search and OptimizationPublished by Wiley ,2003
- Learning inverse kinematicsPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2002
- Using Expectation-Maximization for Reinforcement LearningNeural Computation, 1997
- A Self-Organizing Neural Model of Motor Equivalent Reaching and Tool Use by a Multijoint ArmJournal of Cognitive Neuroscience, 1993
- Forward models: Supervised learning with a distal teacherCognitive Science, 1992
- Learning control for redundant manipulatorsPublished by Institute of Electrical and Electronics Engineers (IEEE) ,1991
- A unified approach for motion and force control of robot manipulators: The operational space formulationIEEE Journal on Robotics and Automation, 1987