AN INTELLIGENT ASSISTANT FOR NAVIGATION OF VISUALLY IMPAIRED PEOPLE N.G. Bourbakis*# and D. Kavraki # #AIIS Inc., Vestal, NY, *WSU,

Slides:



Advertisements
Similar presentations
10 september 2002 A.Broersen Developing a Virtual Piano Playing Environment By combining distributed functionality among independent Agents.
Advertisements

Device receives electronic signal transmitted from signs containing information A device that can communicate GPS location relative to the destination.
Voice Controlled Surgical Assistant ECE 7995 Dupindar ghotra, Muhammad Syed, Sam li, Sophia lei.
Real-time, low-resource corridor reconstruction using a single consumer grade RGB camera is a powerful tool for allowing a fast, inexpensive solution to.
A vision-based system for grasping novel objects in cluttered environments Ashutosh Saxena, Lawson Wong, Morgan Quigley, Andrew Y. Ng 2007 Learning to.
Chapter 5 Input and Output. What Is Input? What is input? p. 166 Fig. 5-1 Next  Input device is any hardware component used to enter data or instructions.
Department of Electrical and Computer Engineering He Zhou Hui Zheng William Mai Xiang Guo Advisor: Professor Patrick Kelly ASLLENGE.
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
INTEGRATION OF A SPATIAL MAPPING SYSTEM USING GPS AND STEREO MACHINE VISION Ta-Te Lin, Wei-Jung Chen, Fu-Ming Lu Department of Bio-Industrial Mechatronics.
CS 561, Sessions 27 1 Towards intelligent machines Thanks to CSCI561, we now know how to… - Search (and play games) - Build a knowledge base using FOL.
44 CHAPTER SPECIALIZED APPLICATION SOFTWARE. © 2005 The McGraw-Hill Companies, Inc. All Rights Reserved. 4-2 Competencies Describe graphics software Discuss.
Supervised by Prof. LYU, Rung Tsong Michael Department of Computer Science & Engineering The Chinese University of Hong Kong Prepared by: Chan Pik Wah,
Experiences with an Architecture for Intelligent Reactive Agents By R. Peter Bonasso, R. James Firby, Erann Gat, David Kortenkamp, David P Miller, Marc.
X96 Autonomous Robot Design Review Saturday, March 13, 2004 By John Budinger Francisco Otibar Scott Ibara.
Final Project CS HCI Kim T Le. Screen Readers for Blind.
Guided Conversational Agents and Knowledge Trees for Natural Language Interfaces to Relational Databases Mr. Majdi Owda, Dr. Zuhair Bandar, Dr. Keeley.
Specialized Application Software
Simultaneous Localization and Map Building System for Prototype Mars Rover CECS 398 Capstone Design I October 24, 2001.
X96 Autonomous Robot Proposal Presentation Monday, February 16, 2004 By John Budinger Francisco Otibar.
Artificial Intelligence
An Integral System for Assisted Mobility Manuel Mazo & the Research group of the SIAMO Project Yuchi Ming, IC LAB.
Rosa Mª Avila Laia Bayarri Cristina Lopera 4r B Ivón Cardenas Maths in English: statistics and its applications Artificial vision.
Introduce about sensor using in Robot NAO Department: FTI-FHO-FPT Presenter: Vu Hoang Dung.
REAL ROBOTS. iCub It has a height of 100 cm, weighs 23 Kg, and is able to recognize and manipulate objects. Each hand has 9 DOF and can feel objects almost.
Track: Speech Technology Kishore Prahallad Assistant Professor, IIIT-Hyderabad 1Winter School, 2010, IIIT-H.
June 12, 2001 Jeong-Su Han An Autonomous Vehicle for People with Motor Disabilities by G. Bourhis, O.Horn, O.Habert and A. Pruski Paper Review.
BlindAid Semester Final Presentation Sandra Mau, Nik Melchior, and Maxim Makatchev.
CIS 601 Fall 2004 Introduction to Computer Vision and Intelligent Systems Longin Jan Latecki Parts are based on lectures of Rolf Lakaemper and David Young.
Kalman filter and SLAM problem
   Input Devices Main Memory Backing Storage PROCESSOR
Computer Vision. DARPA Challenge Seeks Robots To Drive Into Disasters. DARPA's Robotics Challenge offers a $2 million prize if you can build a robot capable.
ASSISTIVE TECHNOLOGY PRESENTED BY ABDUL BARI KP. CONTENTS WHAT IS ASSISTIVE TECHNOLOGY? OUT PUT: Screen magnifier Speech to Recogonizing system Text to.
Specialized Application Software © 2013 The McGraw-Hill Companies, Inc. All rights reserved.Computing Essentials 2013.
Copyright John Wiley & Sons, Inc. Chapter 3 – Interactive Technologies HCI: Developing Effective Organizational Information Systems Dov Te’eni Jane.
Information Technology Industry Report Brown University ADSP Lab 余 渊 善
Multimedia Databases (MMDB)
Group #2 / Embedded Motion Control [5HC99] Embedded Visual Control 1 Group #5 / Embedded Visual Control Self-Balancing Robot Navigation Paul Padila Vivian.
McGraw-Hill/Irwin Copyright © 2008 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 4 Specialized Application Software.
CMPD273 Multimedia System Prepared by Nazrita Ibrahim © UNITEN2002 Multimedia System Characteristic Reference: F. Fluckiger: “Understanding networked multimedia,
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
PMS Software Ltd Electronic Communications A Guide.
Multimedia ITGS. Multimedia Multimedia: Documents that contain information in more than one form: Text Sound Images Video Hypertext: A document or set.
Grenoble Informatics Laboratory (LIG) HCI Engineering Research Group (IIHM) Helping the design of Mixed Systems Céline Coutrix PhD Supervisor: Laurence.
Specialized Input and Output. Inputting Sound ● The microphone is the most basic device for inputting sounds into a computer ● Microphones capture sounds.
MULTIMEDIA INPUT / OUTPUT TECHNOLOGIES INTRODUCTION 6/1/ A.Aruna, Assistant Professor, Faculty of Information Technology.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 4 Specialized Application Software.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. Externally growing self-organizing maps and its application to database visualization and exploration.
Based on the success of image extraction/interpretation technology and advances in control theory, more recent research has focused on the use of a monocular.
Chapter 10. The Explorer System in Cognitive Systems, Christensen et al. Course: Robots Learning from Humans On, Kyoung-Woon Biointelligence Laboratory.
GCAPS Team Design Review CPE 450 Section 1 January 22, 2008 Nick Hebner Kooper Frahm Ryan Weiss.
CONTENT FOCUS FOCUS INTRODUCTION INTRODUCTION COMPONENTS COMPONENTS TYPES OF GESTURES TYPES OF GESTURES ADVANTAGES ADVANTAGES CHALLENGES CHALLENGES REFERENCE.
Ghislain Fouodji Tasse Supervisor: Dr. Karen Bradshaw Computer Science Department Rhodes University 24 March 2009.
It Starts with iGaze: Visual Attention Driven Networking with Smart Glasses It Starts with iGaze: Visual Attention Driven Networking with Smart Glasses.
Projectargus.eu ARGUS Assisting Personal Guidance System for People with Visual Impairment Presented by Manfred Schrenk Managing Director of.
Copyright John Wiley & Sons, Inc. Chapter 3 – Interactive Technologies HCI: Developing Effective Organizational Information Systems Dov Te’eni Jane.
Natural Language and Speech (parts of Chapters 8 & 9)
Presented By: O. Govinda Rao 3 rd MCA AITAM CH. Hari Prasad 3 rd MCA AITAM.
Portable Camera-Based Assistive Text and Product Label Reading From Hand-Held Objects for Blind Persons.
Senior Science Information Systems
How can speech technology be used to help people with disabilities?
Xinguo Yu, Lvzhou Gao, Bin He, Xiaowei Shao May 20, 2017
G. Anushiya Rachel Project Officer
Input and output devices for visually impaired users
Information Computer Technology
Information Computer Technology
CHAITANYA INSTITUTE OF SCIENCE AND TECHNOLOGY
Senior Design Capstone Project I
Instructor: Mike O’Dell
A Mobile Application for the Blind and Dyslexic
Presentation transcript:

AN INTELLIGENT ASSISTANT FOR NAVIGATION OF VISUALLY IMPAIRED PEOPLE N.G. Bourbakis*# and D. Kavraki # #AIIS Inc., Vestal, NY, *WSU, ITRI, Image-Video-Machine Vision Research Lab, Dayton OH.

Outline Introduction Tyflos System Human Tyflos Interaction Visual to Audio Communication Detection of Other Moving Objects Navigation with L-G Landmarks IMPLEMENTATION OF THE TYFLOS PROTOTYPE CONCLUSIONS

Introduction This paper presents the navigation methodology employed by an intelligent assistant (agent) for people with disabilities partial independence. Tyflos –Work in a 3-D dynamic environment –System carries two vision cameras and captures images from the surrounding 3-D environment. –These images into verbal descriptions for each image into a verbal communication.

Introduction In USA –there are more than 3 millions with low vision and millions more completely blind. A great challenge –robotic-like vision Be able to obtain 3-D representations of the surrounding world. provide verbal descriptions of it

Introduction Of particular importance –Unknown environment –High resolution –Analyze –Events based on simple queries –Vision system convert visual descriptions into verbal communication

Introduction Applied artificial intelligence and software technologies –QBIC Query By Image Content –EUROTRA Machine languages translation –STARS Newsletter –MEL An autonomous intelligent vehicle –Intelligent walking robots, 3-D space maps generation, autonomous vision systems, neuromorphic vision chips, the robot with feelings

Introduction verbally see –an intelligent system to make visually impaired people via a machine vision system equipped with speech and natural language understanding technologies. challenges and technological –the navigation methodology of an intelligent assistant for visually impaired people's partial independence.

TYFLOS SYSTEM System's Main Components –The Pair of Glasses Vision Cameras Laser Range Scanner Ear Speaker Microphone Portable Computer Communication System

TYFLOS SYSTEM The Pair of Glasses –The pair of glasses is a important part of the Tyflos system, and several system’s parts are mounted on it.

TYFLOS SYSTEM The Vision Cameras –The vision camera is mounted on the top of the glasses in order to provide a better view of the targeted scene. Camera –High resolution –Color –Small in size and weight.

TYFLOS SYSTEM The Laser Range Scanner –"scan“ the same view with the camera. –For determination of the distances of the existing objects –For 3-D view of a captured image The size of the laser scanner –resolution at least 1cm radius in a distance of 15 meters.

TYFLOS SYSTEM The Ear Speaker –high fidelity component has only one ear speaker through it the user receives the detailed description of the surrounding environment.

TYFLOS SYSTEM The Microphone –User communicates with computer and the appropriate Databases. –Filter To eliminate the noise from the user's environment.

TYFLOS SYSTEM The Portable Computer –It hosts all the important software tools that make the Tyflos system functional. –DBs Image Speech Natural Language –Software Interfaces –Knowledge Conversion Tools –Communication Tools

TYFLOS SYSTEM The Communication System –the user needs to establish communication with other resource portable or mobile information DBs

TYFLOS SYSTEM Source:

HUMAN TYFLOS INTERACTION Tyflos is able to assist the user are : –where am I ? –describe the surrounding scene. –guide me to a particular point –where is a particular object?

HUMAN TYFLOS INTERACTION Independent –Human assistant by describing to the user the 3-D visual environment making him/her more independent. find an “object” → microphone → scanning the surrounding space → vision camera → analyzed and recognized → ear speaker What Object –describes to the user via the ear speaker what object made a certain sound.

Visual to Audio Communication Where am I ? –GPS (Global Positioning System) Tyflos –uses a Speech synthesizer –provides to the user the names of the streets and / or the names of certain buildings. –provides to the human user a verbal description of the surrounding 3-D scene.

Visual to Audio Communication Describe the Surrounding Scene –SPN graph Generation and Description of a 3-D Surrounding Unknown Environment Fusion of Images & range Data for 3-D Modeling of the Free Space

Visual to Audio Communication Generation and Description of a 3-D Surrounding Unknown Environment –stereo speaker –vision camera –laser range data Dr.Albus and Dr. Bourbakis –develop 3-D images via audio signals generated by two speakers during the motion of a sound generation equipment.

Visual to Audio Communication Fusion of Images & range Data for 3-D Modeling of the Free Space –Cameras Stabilization Issues synthesis of images –Generation & Representation of the Navigation Space Free Navigation Space (FNS) –orientation-angle –magnetic north –current position –inclination angle shape SH –” open-navigation corridors ”

Visual to Audio Communication Where is a Particular object? –particular object is converted into a hierarchical SPN graph –the SPN graph is compared to the system's KB –recognized or appropriately classified –speech DB and ear speakers

Visual to Audio Communication Guide me in a Particular Place –Collision Free Path Planning the system attempts to select the First Best Choice (FBC) open navigation corridors. –Detection of Other Moving Objects How Tyflos detects other moving objects in the same navigation space What is the perceived size of the moving object What is the direction and the velocity of the moving object How many different moving objects are in the same free space

Detection of Other Moving Objects How Tyflos detects other moving objects in the same navigation space –the space shape changing methodology shape SH & Dt do(Dt) = VRa*Dt → P' P' → SH' –IF SH" = SH' THEN there is no other moving object in the same navigation space –ELSE there is at least one moving object in the same free navigation space. –the differences detected from the comparison of two consecutive 2-D images. ”motion”

Detection of Other Moving Objects What is the perceived size of the moving object –When Tyflos, detects another moving object, Rb, in the same navigation area, then it attempts to define the size (dimensions) "perceived" from its current position P. Rb = |dk'(t(i+1)) - dk(t(i+1))| –Corresponding window 2-D image Knowledge Base

Navigation with L-G Landmarks L-G Landmarks –local-global graphs

Knowledge Base (KB) –successful and flexible generation of Space Maps is the development of a knowledge base able to store, process and access the Space Maps' forms. Communication and Mobile DBs Issues –to carry CDs with different DBs –its computer to be connected with a multimedia network in order to download the appropriate DB

IMPLEMENTATION OF THE TYFLOS PROTOTYPE Integration of AI Methodologies –The actual implementation of the Tyflos system combines the state-of-the-art in AI methodologies. The Prototype –AIIS (American Institute of Indian Studies) –Tyflos system consists of the pair of glasses, a Hitachi high resolution, color vision camera, a portable computer Mc book- note, an ear speaker, a voice synthesizer and a microphone.

CONCLUSIONS It will provide the ability to the user to understand better the surrounding environment, and give more independence in his/her motion. –Vision –Speech However, a significant progress has been done, and new methods in both vision and speech developed by other researchers will be appropriately used.

Thank you for your attention