Retrieval of blink- based lexicon for persons with brain-stem injury using video cameras CVPR Workshop on Face Processing in Video  June 28, 2004, Washington,

Slides:



Advertisements
Similar presentations
On Facial Recognition in Video VIVA Seminar, U. of Ottawa, August 28, 2003 Dr. Dmitry Gorodnichy Computational Video Group Institute for Information Technology.
Advertisements

ARTIFICIAL PASSENGER.
National Research Council Canada Conseil national de recherches Canada National Research Council Canada Conseil national de recherches Canada Canada Dmitry.
Kien A. Hua Division of Computer Science University of Central Florida.
Face Recognition. Introduction Why we are interested in face recognition? Why we are interested in face recognition? Passport control at terminals in.
Face Detection & Synthesis using 3D Models & OpenCV Learning Bit by Bit Don Miller ITP, Spring 2010.
Reference Guide: Visual feedback provided by Nousor for different states NouseBoard + NousePad Touch corner for next letter selection Go to block containing.
Object Inter-Camera Tracking with non- overlapping views: A new dynamic approach Trevor Montcalm Bubaker Boufama.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
HCI Final Project Robust Real Time Face Detection Paul Viola, Michael Jones, Robust Real-Time Face Detetion, International Journal of Computer Vision,
Motion Detection And Analysis Michael Knowles Tuesday 13 th January 2004.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
MUltimo3-D: a Testbed for Multimodel 3-D PC Presenter: Yi Shi & Saul Rodriguez March 14, 2008.
Face Detection: a Survey Speaker: Mine-Quan Jing National Chiao Tung University.
Cindy Song Sharena Paripatyadar. Use vision for HCI Determine steps necessary to incorporate vision in HCI applications Examine concerns & implications.
CS335 Principles of Multimedia Systems Multimedia and Human Computer Interfaces Hao Jiang Computer Science Department Boston College Nov. 20, 2007.
Stockman MSU Fall Computing Motion from Images Chapter 9 of S&S plus otherwork.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
FACE DETECTION AND RECOGNITION By: Paranjith Singh Lohiya Ravi Babu Lavu.
REAL TIME EYE TRACKING FOR HUMAN COMPUTER INTERFACES Subramanya Amarnag, Raghunandan S. Kumaran and John Gowdy Dept. of Electrical and Computer Engineering,
The Camera Mouse: Visual Tracking of Body Features to Provide Computer Access for People With Severe Disabilities.
Knowledge Systems Lab JN 9/10/2002 Computer Vision: Gesture Recognition from Images Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
Computational Video Group From recognition in brain to recognition in perceptual vision systems. Case study: face in video. Example: identifying computer.
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
Using associative memory principles to enhance perceptual ability of vision systems (Giving a meaning to what you see) CVPR Workshop on Face Processing.
An Information Fusion Approach for Multiview Feature Tracking Esra Ataer-Cansizoglu and Margrit Betke ) Image and.
National Research Council Canada Conseil national de recherches Canada National Research Council Canada Conseil national de recherches Canada Institute.
DIEGO AGUIRRE COMPUTER VISION INTRODUCTION 1. QUESTION What is Computer Vision? 2.
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Access Control Via Face Recognition. Group Members  Thilanka Priyankara  Vimalaharan Paskarasundaram  Manosha Silva  Dinusha Perera.
Face Recognition in Video Int. Conf. on Audio- and Video-Based Biometric Person Authentication (AVBPA ’03) Guildford, UK June 9-11, 2003 Dr. Dmitry Gorodnichy.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Tracking CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
A study on face system Speaker: Mine-Quan Jing National Chiao Tung University.
Recognizing faces in video: Problems and Solutions NATO Workshop on "Enhancing Information Systems Security through Biometrics" October 19, 2004 Dr. Dmitry.
Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais.
Optical flow and keypoint tracking Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Face recognition in video as a new biometrics modality & the associative memory framework Talk for IEEE Computation Intelligence Society (Ottawa Chapter)
Over the recent years, computer vision has started to play a significant role in the Human Computer Interaction (HCI). With efficient object tracking.
Coin Recognition Using MATLAB - Emad Zaben - Bakir Hasanein - Mohammed Omar.
Presented By Bhargav (08BQ1A0435).  Images play an important role in todays information because A single image represents a thousand words.  Google's.
National Taiwan Normal A System to Detect Complex Motion of Nearby Vehicles on Freeways C. Y. Fang Department of Information.
Course : T Computer Vision
Hand Gestures Based Applications
Presented by Jason Moore
EYE-GAZE COMMUNICATION
Neural Networks.
Gait Recognition Gökhan ŞENGÜL.
by Konrad Rzeszutek B.S. University of New Orleans, 1999
CS201 Lecture 02 Computer Vision: Image Formation and Basic Techniques
Tracking Objects with Dynamics
Motion Detection And Analysis
FACE DETECTION USING ARTIFICIAL INTELLIGENCE
Seunghui Cha1, Wookhyun Kim1
Common Classification Tasks
Efficient Deformable Template Matching for Face Tracking
On Facial Recognition in Video
The Implementation of a Glove-Based User Interface
Introduction to Artificial Intelligence Lecture 24: Computer Vision IV
AHED Automatic Human Emotion Detection
Fast Forward, Part II Multi-view Geometry Stereo Ego-Motion
Optical flow and keypoint tracking
Towards building user seeing computers
Presentation transcript:

Retrieval of blink- based lexicon for persons with brain-stem injury using video cameras CVPR Workshop on Face Processing in Video  June 28, 2004, Washington, DC, USA Dr. Dmitry Gorodnichy Computational Video Group Institute for Information Technology National Research Council Canada www.cv.iit.nrc.ca/Nouse www.perceptual-vision.com

This work summarizes the problems and solutions Some people cannot use hands to control computer. For some, blinking is the only way to communicate… Types of reply: (to “Do you need anything?”) “YES” – blink “NO” – no blink Non-binary replies are possible using various types of blinks Main solution so far – EMG (Electromyography) readings - Electrodes are attached to skin around oblicularis oculi muscle used in eye closure. - Not effective, because of electrical noise due to skin conductance changes and slight positioning changes of the electrodes. What computer vision can offer? – Perceptual User Interfaces 1. Using high-resolution video [fpiv04-ey1] – expensive, slow 2. Using structured infrared [fpiv04-ft4] – intrusive, problems with sun-light, eye-glasses 3. Using motion component of video This is how it is done by biological vision systems: people can easily detect a blink in video with faces being as small as 20 pixels between the eyes! Not simple for computer vision systems & not much work seems to be done. This work summarizes the problems and solutions “YES” – two blinks “NO” – three blinks “YES” – long blink “NO” – short blink

Perceptual User Interfaces Internet, tendencias & tecnología La nariz utilizada como mouse En el Instituto de Tecnología de la Información, en Canadá, se desarrolló un sistema llamado Nouse que permite manejar softwares con movimientos del rostro. El creador de este programa, Dmitry Gorodnichy, explicó vía e-mail a LA NACION LINE cómo funciona y cuáles son sus utilidades Si desea acceder a más información, contenidos relacionados, material audiovisual y opiniones de nuestros lectores ingrese en : http://www.lanacion.com.ar/03/05/21/dg_497588.asp Copyright S. A. LA NACION 2003. Todos los derechos reservados. x y , z a , b g PUI monitor binary event ON OFF recognition / memorization Unknown User! Goal: to detect, track and recognize face and facial movements of the user.

Evolution of Perceptual User Interfaces 1998. Proof-of-concept colour-based skin tracking [Bradski’98] – not precise though 2001. Motion-based segmentation & localization – not precise though 1999-2002. Several skin colour models developed - unlikely to get much better than that 2001. Viola-Jones face detection using Haar-like wavelets of intensities [fg02] 2002. Subpixel-accuracy convex-shape nose tracking [Nouse, fg02, ivc04] 2002. Stereo face tracking using projective vision [w.Roth, ivc04] 2003. Second-order change detection [Double-blink, ivc04] 2004. Associative visual memory for localized faces [fpiv04-fa2] Figure: Typical result for face detection using colour, motion and intensity components of video using six different webcams.

Nouse™ perceptual vision technology Precision & convenience of tracking the convex-shape nose feature is such that it allows one to use nose as mouse (or joystick handle) [fg02] Next step taken: for people with normal eye lid motion for both eyes - Use eye blinking to replace mouse click Next step to be taken: for people with disabilities who are able to eye blink (wink) only - Reading eye-blink based lexicon

Detection of “Blinked” commands Overview (using motion component of video): stream of video data: I(t) three video frames: I1, I2, I3 change maps: binary image(s) time-stamped “blink” event “blink”-ed command 0: 1: 2: 3: Stages: 0: Detecting a face (eyes) - optional 1: Detecting changes cause by motion 2. Analyzing the detected changes 3. Combining results over time Constraint: All has to be done in real-time (<100ms) Setup: Video: 160x120 resolution suffices!

How to detect change Difficulty with: 1. moving heads (change around other facial features occurs)  second-order change detection can help 2. changing light conditions  non-linear change detection can help [Durucan’02] 3. due to eye closing vs due eye opening examine both negative and positive change imagess 4. one-eye blinks (winks): lack of spatial constraints, other parts move too describe the change caused by a wink using moments compute moments in overlapping regions: around eye & around face Example: changes caused by right eye blink (right image) can be very similar to those caused by slight head motion (left image)

Second-order change detection Key idea: to separate long lasting change from the most recent change (i.e. to detect change in change) dI’ V It-2 It It-1 dI dI’(shifted by V) - * - changed pixels (1) - unchanged pixels(0) ddI 1: 2: Examples: 1. Closing the protector, while moving a diskette 2. Blinking, while moving a head

How to analyze change images Using second-order change to compute eye closings and openings I I1 I2 _ dIon dIoff + dI dI1 = dIlong ddI dI1(V) & Long-term change dIlong = dI – ddI (not shown) can be used to refine positive and negative change images dIon, dIoff Using moments to recognize blobs of predefined shapes: Computing moments is not robust to outliers:  Outliers can be removed by second order change detection  Computing moments in overlapping regions helps too.

Implementation example using PVS.dll For two-eye blinks: '''initialize camera Sub FormLoad() pvsInitializeCamera(160, 120, 1) nIOD = 20 End Sub ''' define change caused by blink Function EyeLidMoved() As Boolean ... End Function ''' for each captured frame Sub Timer_Timer() pvsInitializeNextFrame() nTimestamp = pvsGetFrameNumber() pvsComputeChanges() ' detect eye closings by computing N,X,Y,dX,dY,tilt in negative change If EyeLidMoved( pvsAnalyzeChange(roi1,dIoff), pvsAnalyzeChange(roi2,dIoff) ) Then addToList(timeline,bClosing,nTimestamp) ' detect eye openings by computing N,X,Y,dX,dY,tilt in positive change If EyeLidMoved( pvsAnalyzeChange(roi1,dIon), pvsAnalyzeChange(roi2,dIon) ) Then add2List(timeline,bOpening,nTimestamp) ' detect blink by analyzing time-line signatures in the last K time-stamped eye lid motions If isThereBlink(timeline, K) Then ... 'Blink detected! EyeLidMoved() { if (r<N<r*iod & r<dX<2*iod & dY<r & tilt<1/4) return true; } // iod is intra-ocular distance, r=iod/4. For one-eye blinks (winks): - additional constraints are needed: on eye / face location (Xe,Ye) - this can be obtained using Viola-Jones face detector.

Associative memory for face analysis Potential approach: - store facial expressions as attractors in the attractor-based neural network. - such network of N binary neurons Yi, converges in time from a stimulus binary visual pattern to an attractor according to the update rule: Main question: How to compute synaptic weights Cij so that the network exhibits the best associative recall? The answer in [Gorodnichy-fpiv04-fa04]

Conluding remarks Techniques to be used for eye blink detection are defined. However, each case should be approached individually, depending on properties of the individual's eye lid movement, and computer power available. Time vs reliability tradeoff: The amount of processing should permit real-time execution (<100msec) so that blinks are not missed. On the other hand, less processing, worse change detection. Processing time needed for different video processing tasks (for 160x120 video on 1GHz Celeron laptop, in msec) * computes change based on the intensity in the neighbourhood of a pixel rather than on the intensity of the pixel itself . ** performed with an attractor-based neuron network of 24x24 neurons Acknowledgement: this technology is currently under test at the Midlands Rehabilitation Center, UK no processing 10 Colour-based detection 50 Motion-based detection 40 Viola-Jones face detection 140 Associative recall ** 100 maximal expected motion (pixels) 0 1 2 3 4 5 Second-order change detection 40 80 100 140 180 230 Non-linear change detection* 250