Update September 14, 2011 Adrian Fletcher, Jacob Schreiver, Justin Clark, & Nathan Armentrout
Progress ◦ Integrated Development Environment & Toolkit Setup ◦ Video/Image Acquisition & Image Processing ◦ Subversion, Website, and Documentation Proposed Approach Goals ◦ Image Segmentation by Sept. 21 ◦ Camera Calibration and Parameterization Software by Sept. 21 ◦ SyRS base-line revision complete by Sept. 28 Questions
Development Software Decided ◦ NetBeans with OpenCV through JavaCV wrapper Works on every team member's computer Netbeans is a free IDE Platform Independent* Document made that explains how to install all software components Edge detection and contour interpolation program created for testing ◦ Shows contours as blue lines in real time ◦ Utilizes video source from built in webcam
5 Major Components ◦ Video/Image Acquisition & Enhancement ◦ Image Segmentation ◦ Image Registration using Optical Flow Analysis ◦ 3-D Projection Mapping from Camera Parameterization ◦ Provide Directional Output for Navigation
Canny Edge Detection on Webcam video Contour Interpolation
Website made with auto-generated content based solely on SVN ◦ Automatic revision tracking ◦ Automatic update notifications ◦ Updates on an hourly basis Project description written and submitted to Dr. Elmaghraby SyRS started with project description and background information (needs additional input)
Image Processing ◦ Analyze image segmentation techniques and begin coding at minimum ◦ Analyze approaches for registration using optical flow Camera Calibration and Parameterization ◦ Utilize ProCamLib software tool that is suggested by JavaCV (if possible) or select an appropriate tool ◦ Find the camera parameterization matrix (of webcam) Documentation ◦ Finish SyRS initial base-line
What are the camera specifications & parameters? What sensors are on the MAV and do we have access to sensors? Can we use these sensors to get information about the velocity and angular velocity (most important)? Suggestions?