Reinforcement learning of humanoid rhythmic walking parameters based on visual information
- 1 January 2004
- journal article
- Published by Taylor & Francis in Advanced Robotics
- Vol. 18 (7) , 677-697
- https://doi.org/10.1163/1568553041719519
Abstract
This paper presents a method for learning the parameters of rhythmic walking to generate purposive humanoid motions. The controller consists of the two layers: rhythmic walking is realized by the lower layer, which adjusts the speed of the phase on the desired trajectory depending on sensory information, and the upper layer learns (i) the feasible parameter sets that enable stable walking, (ii) the causal relationship between the walking parameters to be given to the lower-layer controller and the change in the sensory information and (iii) the feasible rhythmic walking parameters by reinforcement learning so that a robot can reach the goal based on visual information. The experimental results show that a real humanoid learns to reach the ball and to shoot it into the goal in the context of the RoboCup soccer competition, and the further issues are discussed.Keywords
This publication has 6 references indexed in Scilit:
- A model of the neuro-musculo-skeletal system for anticipatory adjustment of human locomotion during obstacle avoidanceBiological Cybernetics, 1998
- The RoboCup humanoid challenge as the millennium challenge for advanced roboticsAdvanced Robotics, 1998
- Reinforcement LearningPublished by Springer Nature ,1992
- Self-organized control of bipedal locomotion by neural oscillators in unpredictable environmentBiological Cybernetics, 1991
- The Role of Visual Information in Control of a Constrained Locomotor TaskJournal of Motor Behavior, 1988
- Neurobiological Bases of Rhythmic Motor Acts in VertebratesScience, 1985