Standing Balance Control Using a Trajectory Library

Slides:



Advertisements
Similar presentations
Viktor Zhumatiya, Faustino Gomeza,
Advertisements

State Space Models. Let { x t :t T} and { y t :t T} denote two vector valued time series that satisfy the system of equations: y t = A t x t + v t (The.
Parallelizing Video Transcoding With Load Balancing On Cloud Computing Song Lin, Xinfeng Zhang, Qin Y, Siwei Ma Circuits and Systems, 2013 IEEE.
Jose-Luis Blanco, Javier González, Juan-Antonio Fernández-Madrigal University of Málaga (Spain) Dpt. of System Engineering and Automation May Pasadena,
Tracking Unknown Dynamics - Combined State and Parameter Estimation Tracking Unknown Dynamics - Combined State and Parameter Estimation Presenters: Hongwei.
« هو اللطیف » By : Atefe Malek. khatabi Spring 90.
1 Balancing Push and Pull for Efficient Information Discovery in Large-Scale Sensor Networks Xin Liu, Qingfeng Huang, Ying Zhang CS 6204 Adv Top. in Systems-Mob.
Manifold Sparse Beamforming
Low Complexity Keypoint Recognition and Pose Estimation Vincent Lepetit.
Model base human pose tracking. Papers Real-Time Human Pose Tracking from Range Data Simultaneous Shape and Pose Adaption of Articulated Models using.
Optimization methods Morten Nielsen Department of Systems biology, DTU.
Introduction to Mobile Robotics Bayes Filter Implementations Gaussian filters.
Active SLAM in Structured Environments Cindy Leung, Shoudong Huang and Gamini Dissanayake Presented by: Arvind Pereira for the CS-599 – Sequential Decision.
Using Inaccurate Models in Reinforcement Learning Pieter Abbeel, Morgan Quigley and Andrew Y. Ng Stanford University.
XYZ 6/18/2015 MIT Brain and Cognitive Sciences Convergence Analysis of Reinforcement Learning Agents Srinivas Turaga th March, 2004.
Discriminative Training of Kalman Filters P. Abbeel, A. Coates, M
Prepared By: Kevin Meier Alok Desai
7. Experiments 6. Theoretical Guarantees Let the local policy improvement algorithm be policy gradient. Notes: These assumptions are insufficient to give.
Source-Channel Prediction in Error Resilient Video Coding Hua Yang and Kenneth Rose Signal Compression Laboratory ECE Department University of California,
Markov Decision Processes Value Iteration Pieter Abbeel UC Berkeley EECS TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Reinforcement Learning Game playing: So far, we have told the agent the value of a given board position. How can agent learn which positions are important?
Advanced Topics in Optimization
Motion Planning in Dynamic Environments Two Challenges for Optimal Path planning.
Solving Problems by Searching AIMA Chapter 3 Fall 2006 Messiah College Dr. Gene Chase.
Digital Filter Stepsize Control in DASPK and its Effect on Control Optimization Performance Kirsten Meeker University of California, Santa Barbara.
CHEN RONG ZHANG WEN JUN.  Introduction and Features  Price model including Bayesian update  Optimal trading strategies  Coding  Difficulties and.
Overview and Mathematics Bjoern Griesbach
Real-Time Odor Classification Through Sequential Bayesian Filtering Javier G. Monroy Javier Gonzalez-Jimenez
EKF and UKF Day EKF and RoboCup Soccer simulation of localization using EKF and 6 landmarks (with known correspondences) robot travels in a circular.
Markov Localization & Bayes Filtering
Target Tracking with Binary Proximity Sensors: Fundamental Limits, Minimal Descriptions, and Algorithms N. Shrivastava, R. Mudumbai, U. Madhow, and S.
Optimal Nonlinear Neural Network Controllers for Aircraft Joint University Program Meeting October 10, 2001 Nilesh V. Kulkarni Advisors Prof. Minh Q. Phan.
Space-Indexed Dynamic Programming: Learning to Follow Trajectories J. Zico Kolter, Adam Coates, Andrew Y. Ng, Yi Gu, Charles DuHadway Computer Science.
Simultaneous Localization and Mapping Presented by Lihan He Apr. 21, 2006.
IEEE Globecom 2010 Tan Le Yong Liu Department of Electrical and Computer Engineering Polytechnic Institute of NYU Opportunistic Overlay Multicast in Wireless.
Kalman Filter (Thu) Joon Shik Kim Computational Models of Intelligence.
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
Particle Filters.
Remarks: 1.When Newton’s method is implemented has second order information while Gauss-Newton use only first order information. 2.The only differences.
Approximate Dynamic Programming Methods for Resource Constrained Sensor Management John W. Fisher III, Jason L. Williams and Alan S. Willsky MIT CSAIL.
1 Distributed and Optimal Motion Planning for Multiple Mobile Robots Yi Guo and Lynne Parker Center for Engineering Science Advanced Research Computer.
Solving POMDPs through Macro Decomposition
Akram Bitar and Larry Manevitz Department of Computer Science
Maximum a posteriori sequence estimation using Monte Carlo particle filters S. J. Godsill, A. Doucet, and M. West Annals of the Institute of Statistical.
Sparse Signals Reconstruction Via Adaptive Iterative Greedy Algorithm Ahmed Aziz, Ahmed Salim, Walid Osamy Presenter : 張庭豪 International Journal of Computer.
Combs, Needles, Haystacks: Balancing Push and Pull for Discovery in Large Scale Sensor Networks Xin Liu Department of Computer Science University of California.
Håkon Dahl-Olsen, Sridharakumar Narasimhan and Sigurd Skogestad Optimal output selection for batch processes.
Multi-Speaker Modeling with Shared Prior Distributions and Model Structures for Bayesian Speech Synthesis Kei Hashimoto, Yoshihiko Nankaku, and Keiichi.
Mixture Kalman Filters by Rong Chen & Jun Liu Presented by Yusong Miao Dec. 10, 2003.
A Dynamic Query-tree Energy Balancing Protocol for Sensor Networks H. Yang, F. Ye, and B. Sikdar Department of Electrical, Computer and systems Engineering.
Bayesian Speech Synthesis Framework Integrating Training and Synthesis Processes Kei Hashimoto, Yoshihiko Nankaku, and Keiichi Tokuda Nagoya Institute.
By: Aaron Dyreson Supervising Professor: Dr. Ioannis Schizas
Attribute Allocation in Large Scale Sensor Networks Ratnabali Biswas, Kaushik Chowdhury, and Dharma P. Agrawal International Workshop on Data Management.
Repairing Sensor Network Using Mobile Robots Y. Mei, C. Xian, S. Das, Y. C. Hu and Y. H. Lu Purdue University, West Lafayette ICDCS 2006 Speaker : Shih-Yun.
Zhaoxia Fu, Yan Han Measurement Volume 45, Issue 4, May 2012, Pages 650–655 Reporter: Jing-Siang, Chen.
Web Servers load balancing with adjusted health-check time slot.
Robust and Fast Collaborative Tracking with Two Stage Sparse Optimization Authors: Baiyang Liu, Lin Yang, Junzhou Huang, Peter Meer, Leiguang Gong and.
System Control based Renewable Energy Resources in Smart Grid Consumer
Privacy and Fault-Tolerance in Distributed Optimization Nitin Vaidya University of Illinois at Urbana-Champaign.
Student: Hao Xu, ECE Department
Design Space Exploration
Lecture 10: Observers and Kalman Filters
Robust Control Chris Atkeson 5/3/16.
Bowei Tang, Tianyu Chen, and Christopher Atkeson
Bursty and Hierarchical Structure in Streams
Alfred Lynam February 21, 2008 Dynamics and Control D&C simulation code compatibility test with trajectory code. AAE 450 Spring 2008 Dynamics and Control.
3-Dimensional Structure
October 20, 2010 Dr. Itamar Arel College of Engineering
Akram Bitar and Larry Manevitz Department of Computer Science
Outlines Introduction & Objectives Methodology & Workflow
Presentation transcript:

Standing Balance Control Using a Trajectory Library Chenggang Liu and Chris Atkeson

Outline Introduction Robot Model Neighboring Optimal Control Balance Controller Simulation and Experiment Results

Introduction Multiple Balance Strategies from One Optimization Criterion.

Robot Model

Trajectory Library Generation Combined Method Direct minimization with SNOPT (Sequential Quadratic Programming) Differential Dynamic Programming A library on a uniform grid of initial conditions

Neighboring Optimal Control Given the discrete time dynamics of the robot: and the optimal value function: The neighboring optimal control is given by: Where

Neighboring Optimal Control Having the optimal trajectories of state over time, , control over time, , and the gain matrices over time, . A local approximation to optimal control is: where is the closest state to the current state on , and are those corresponding to on and .

Controller Architecture

Trajectory Library Generation The library is refined to get a new library on an adaptive grid of initial conditions according to the proposed controller’s performance.

State/Push Estimation State to be estimated: Measurements:

State/Push Estimation State transition and observation models

State/Push Estimation To predict the next state before measurements are taken: To update the state after measurements are taken:

Simulation Results Constant push of 42 Newtons at head Short forward push at head of 50 Newtons, lasting 0.5 seconds Random pushes sequence

Simulation Results Comparison with the optimal controller.

Experiment Result Push forward Trajectory index State index

Experiment Result Push backward Trajectory index State index

Thanks! For more information: http://www.andrew.cmu.edu/user/cgliu/CurrentRD.html