Download presentation
Published byJeremy Russell Modified over 9 years ago
1
Function Approximation for Imitation Learning in Humanoid Robots Rajesh P. N. Rao Dept of Computer Science and Engineering University of Washington, Seattle neural.cs.washington.edu Students: Rawichote Chalodhorn, David Grimes Funding: ONR, NSF, Packard Foundation
2
The Problem: Robotic Imitation of Human Actions
TODO: 1) Teacher (David Grimes) HOAP-2 Humanoid Robot (Morpheus or Mo)
3
Example of Motion Capture Data
Motion Capture Sequence Attempted Imitation
4
Goals Learn from only observations of teacher states
Expert does not control robot Also called “implicit imitation” (Price & Boutilier, 1999) Similar to how humans learn from imitation Avoid hand-coded physics-based models Learn dynamics in terms of sensory consequences of executed actions Use teacher demonstration to restrict search space of feasible actions
5
Step 1: Kinematic mapping
Need to solve the “correspondence problem” Solved by assuming markers are on scaled version of robot body Standard inverse kinematics recovers joint angles for motion
6
Step 2: Dimensionality Reduction
Humanoid robots have large DOF, making action optimization intractable HOAP-2 has 25 DOF Fortunately, most actions are highly redundant Can use dimensionality reduction techniques (e.g., PCA) to represent states and actions
7
Posture Representation using Eigenposes
8
Eigenposes for Walking
9
Step 3: Learning Forward Models using Function Approximation
Basic Idea: 1. Learn forward model in the neighborhood of teacher demonstration Use function approximation techniques to map actions to observed sensory consequences 2. Use the learned model to infer stable actions for imitation 3. Iterate between 1 and 2 for higher accuracy
10
Approach 1: RBF Networks for Deterministic Action Selection
Radial Basis Function (RBF) network used to learn the n-th order Markov function: st is the sensory state vector E.g., st = t (3D gyroscope signal) at is the action vector in latent space E.g., Servo joint angle commands in latent space
11
Action Selection using the Learned Function
Select optimal action for next time step t: measures torso stability based on predicted gyroscope signals: Search for optimal action at* limited to local region around teacher trajectory in subspace (Chalodhorn et al., Humanoids, 2005; IJCAI 2007; IROS, 2009)
12
Example: Learning to Walk
Human motion capture data Unoptimized (kinematic) imitation
13
Example: Learning to Walk
Motion scaling Take baby steps first (literally!) Final Result (Chalodhorn et al., IJCAI 2007)
14
Result: Learning to Walk
Human Motion Capture Optimized Stable Walk
15
Approach 2: Gaussian Processes for Probabilistic Action Selection
Dynamic Bayesian Network (DBN) for Imitation [Slice at time t] Ot are observations of states St St = low-D joint space, gyro, foot pressure readings Ct are constraints on states (e.g., gyroscope values near zero) (Grimes et al., RSS 2006; NIPS 2007; IROS 2007; IROS 2008)
16
DBN for Imitative Learning
Gaussian Process-based Forward Model (input [st-1,at]): (Grimes, Chalodhorn, & Rao, RSS 2006)
17
Action Inference using Nonparametric Belief Propagation
Maximum marginal posterior actions Evidence (blue nodes)
18
Summary of Approach Learning and action inference are interleaved to yield progressively more accurate forward models and actions
19
Example of Learning
20
Progression of Imitative Learning
21
Result after Learning Human Action Imitation
(Grimes, Rashid, & Rao, NIPS 2007)
22
Other Examples
23
From Planning to Policy Learning
Behaviors shown in the previous slides were open-loop, based on planning by inference Can we learn closed-loop “reactive” behaviors? Idea: Learn state-to-action mappings (“policies”) based on the final optimized output of the planner and resulting sensory measurements
24
Policy Learning using Gaussian Processes
For a parameterized task T(), watch demonstrations for particular values of E.g., Teacher lifting objects of different weight Parameter not given but intrinsically encoded in sensory measurements Use inference-based planning to infer stable actions at and states st for demonstrated values of Learn Gaussian process policy based on {st, at}: (Grimes & Rao, IROS 2008)
25
Example: Learning to Lift Objects of Different Weights
26
Generalization by Gaussian Process Policy
27
Generalizing to a Novel Object
28
Summary and Conclusions
Stable full-body human imitation in a humanoid robot may be achievable without a physics-based model Function approximation techniques play a crucial role in learning a forward model and in action inference RBF networks, Gaussian processes Function approximation also used to learn policies for reactive behavior Dimensionality reduction using PCA (via “eigenposes”) helps keep learning and inference tractable Challenges: Scaling up to large number of actions, smooth transition between actions, hierarchical control
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.