Presentation is loading. Please wait.

Presentation is loading. Please wait.

CH24 in Robotics Handbook Presented by Wen Li Ph.D. student Texas A&M University.

Similar presentations


Presentation on theme: "CH24 in Robotics Handbook Presented by Wen Li Ph.D. student Texas A&M University."— Presentation transcript:

1 CH24 in Robotics Handbook Presented by Wen Li Ph.D. student Texas A&M University

2  Visual Servo Control  Image based visual servo  Position based visual servo  Hybrid visual servo and other issues  Target Tracking

3  Visual Servo Control  Image based visual servo  Position based visual servo  Hybrid visual servo and other issues  Target Tracking

4  Vision Based Robot Control Task:  USE - computer vision data  CONTROL - motion of a robot

5  Camera Configuration: Eye-in-hand Fixed in workspace

6 Servoing Architecture

7  Basic Components  Image features  Error function  Velocity controller  Interaction matrix

8  Basic Components  Image features  Error function  Velocity controller  Interaction matrix s(m(t),a) ; a is a set of parameters that represent potential additional knowledge about the system (e.g. Camera intrinsic parameters); m(t) is a set of image measurements (e.g. Image coordinates of interest points) s* contains the desired values of the features.

9

10  Basic Components  Image features  Error function  Velocity controller  Interaction matrix e(t)=s(m(t),a)-s* The aim of the control scheme is to minimize error e(t) At the desired pose, e(t)= 0.

11  Basic Components  Image features  Error function  Velocity controller  Interaction matrix The control law v c – the spatial velocity of the camera, input to the robot controller Problem: what is the form of Ls

12  Basic Components  Image features  Error function  Velocity controller  Interaction matrix L s is the interaction matrix, which describes the relationship between the time variation of s and the camera velocity v c., Le=Ls is the approximation of the pseudo-inverse of L s. Problem: how to estimate -- according to different designs of s

13  Categories:  Image based control  Position based control

14  Visual Servo Control  Image based servo control  Position based servo control  Hybrid visual servo and other issues  Target Tracking

15 Ls S(m(t),a)

16

17  Image features s(m(t),a)  Traditionally, s is defined by the image-plane coordinates of a set of points. s=x=(x,y) (x,y)

18  Interaction Matrix The value Z is the depth of the point relative to the camera frame. Therefore, any control scheme that uses this form of the interaction matrix must estimate or approximate the value of Z. When Z is not known, cannot be directly used. An approximation must be used. To control six degrees of freedom, at least three points are necessary. There exists some configurations for which Lx is singular.

19  Effects of different estimations of Ls

20  Advantages:  The positioning accuracy of the system is less sensitive to camera calibration errors  Computational advantage  Disadvantages:  Presence of singularity  Servoing in 2-D

21  Visual Servo Control  Image based servo control  Position based servo control  Hybrid visual servo and other issues  Target Tracking

22 Ls S

23  extract the image features ->  compute the current camera pose with respect to a reference coordinate on the object ->  compare with the desired camera pose with respect to the reference coordinate on the object Current pose desired pose x y z

24  Consider three coordinate frames:  The current camera frame  The desired camera frame  A reference frame attached to the object  gives the coordinates of the origin of the object frame to the current camera frame  gives the coordinates of the origin of the object frame to the desired camera frame , the rotation matrix that gives the orientation of the current camera frame relative to the desired frame

25 Current pose desired pose o

26  Define s=(t,θu)  t is a translation vector, θu is the angle/axis parameterization for the rotation  1) t is defined relative to the object frame

27  Define s=(t,θu)  t is a translation vector, θu is the angle/axis parameterization for the rotation  2) t is defined relative to the desired camera frame

28 Effects of different designs

29  Advantages:  Possible to describe tasks in terms Cartesian pose as is common in Robotics  Disadvantages:  Sensitive to calibration error  Depend on having an accurate mode of target objects – a form of calibrations  Servoing in 3-D

30  Visual Servo Control  Image based servo control  Position based servo control  Hybrid servo and other issues  Target Tracking

31  Hybrid VS – combining 2-D and 3-D features  2.5-D visual servo – add depth of the point  s ▪ Camera trajectory is a straight line ▪ Image trajectory of the center of the gravity of the object is also a straight line

32  Stereo vision system in IBVS Because of epipolar constraint, this approach actually requires 3-D parameters in s. Thus, it would be, strictly speaking, a position-based approach

33  Visual Servo Control  Image based servo control  Position based servo control  Hybrid visual servo and other issues  Target Tracking

34  Moving target => varying value s*(t) Moving target The time variation of e due to the generally unknown target motion Estimate ∂e/∂tImprove estimated value using Kalman filter or more-elaborate filtering methods

35


Download ppt "CH24 in Robotics Handbook Presented by Wen Li Ph.D. student Texas A&M University."

Similar presentations


Ads by Google