MTA SZTAKI An assistive interpreter tool using glove-based hand gesture recognition Department of Distributed Systems Péter Mátételki Máté Pataki Sándor.

Slides:



Advertisements
Similar presentations
Technical and design issues in implementation Dr. Mohamed Ally Director and Professor Centre for Distance Education Athabasca University Canada New Zealand.
Advertisements

Breakout session B questions. Research directions/areas Multi-modal perception cognition and interaction Learning, adaptation and imitation Design and.
Hand Gesture for Taking Self Portrait Shaowei Chu and Jiro Tanaka University of Tsukuba Japan 12th July 15 minutes talk.
Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
Arabic Sign Language Recognition Mohamed Mohandes King Fahd University of Petroleum and Minerals
Interaction Devices By: Michael Huffman Kristen Spivey.
Shweta Jain 1. Motivation ProMOTE Introduction Architectural Choices Pro-MOTE system Architecture Experimentation Conclusion and Future Work Acknowledgement.
PRINVESTIGTOR Dr JYOTI TANDUKAR PRESENTED BY: Er. VIVEK RAJ SHRESTHA Er. PRABHAT ADHIKARI Er. DEVENDRA KATHAYAT.
Hands-free Control of Standard DICOM Imaging Software using Leap Motion or Myo Controller. Peter Stoll & José Morey As Shown at RSNA 2014.
Department of Electrical and Computer Engineering He Zhou Hui Zheng William Mai Xiang Guo Advisor: Professor Patrick Kelly ASLLENGE.
Deborah Gilden, Ph.D. Rehabilitation Engineering Research Center The Smith-Kettlewell Eye Research Institute.
VADE - Virtual Assembly Design Environment Virtual Reality & Computer Integrated Manufacturing Lab.
Gyration GyroMouse. Digitizers 3D Digitization Data Gloves (Haptic Devices)
Department of Electrical and Computer Engineering He Zhou Hui Zheng William Mai Xiang Guo Advisor: Professor Patrick Kelly ASLLENGE Final Project Review.
PROJECT UPDATE:DECEMBER 2014 Yicun Xiong #
Personalized Medicine Research at the University of Rochester Henry Kautz Department of Computer Science.
HAND GESTURE BASED HUMAN COMPUTER INTERACTION. Hand Gesture Based Applications –Computer Interface A 2D/3D input device (Hand Tracking) Translation of.
Final Presentation. Lale AkarunOya Aran Alexey Karpov Milos Zeleny Hasim Sak Erinc Dikici Alp Kindiroglu Marek Hruz Pavel Campr Daniel Schorno Alexander.
Current Technology Trends & Challenges in the Classroom: from BYOD to One Size Fits All Jacqueline Hess Director, Family Center on Technology & Disability.
Project By: Brent Elder, Mike Holovka, Hisham Algadaibi.
Prepared by Group 1 James Monaghan Travis Chandler John Peters Markin Williamson Prepared for Final Presentation Advisor Professor Morse.
Visual and Specialised Interfaces. 3D Software Developments have led to visual systems such as pictures of the earth’s structure and reproductions of.
Kim Kasee Motion Computing. Mobile Worker Issues Mobile computing is a requirement Usage is limited by clamshell designs –Displays are viewed as barriers.
Evolution of Web Accessibility Meenakshi Sripal COMS E6125.
MTA SZTAKI Department of Distributed Systems The problems of persistent identifiers in the context of the National Digital Data Archives of Hungary András.
出處: Signal Processing and Communications Applications, 2006 IEEE 作者: Asanterabi Malima, Erol Ozgur, and Miijdat Cetin 2015/10/251 指導教授:張財榮 學生:陳建宏 學號: M97G0209.
Real Time Appearance Based Hand Tracking The 19th International Conference on Pattern Recognition (ICPR) December 7-11, 2008, Tampa Convention Center,
ECE 8443 – Pattern Recognition EE 3512 – Signals: Continuous and Discrete Objectives: Spectrograms Revisited Feature Extraction Filter Bank Analysis EEG.
Microsoft Assistive Technology Products Brought to you by... Jill Hartman.
Model of the Human  Name Stan  Emotion Happy  Command Watch me  Face Location (x,y,z) = (122, 34, 205)  Hand Locations (x,y,z) = (85, -10, 175) (x,y,z)
HCI 입문 Graphics Korea University HCI System 2005 년 2 학기 김 창 헌.
Sean M. Ficht.  Problem Definition  Previous Work  Methods & Theory  Results.
Hand Motion Identification Using Independent Component Analysis of Data Glove and Multichannel Surface EMG Pei-Jarn Chen, Ming-Wen Chang, and and Yi-Chun.
Supervisor: Dr. Elsayed Eissa Hemayed. o Marwa Ibrahim Lamey. Mayada Ibrahim Aly. o Mona Sherif Ahmed. o Suad Mohamed Barakat. o Marwa Ibrahim Lamey.
GLOBAL INITIATIVE FOR INCLUSIVE ICTs Promoting the Rights of Persons with Disabilities in the Digital Age Leveraging Mobile For Accessible.
Collaborator Revolutionizing the way you communicate and understand
A SSISTIVE TECHNOLOGY TOOLS Morgan McGlamery EDN 303.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais.
BatShot. Face Detection Team BatShot Camera Made Easy Image Correction Tap and Talk Audio and vibration feedback to capture photo Adam Halem Nilesh Chaturvedi.
Giri.K.R [4jn08ec016] Harish.Kenchangowdar[4jn10ec401] Sandesh.S[4jn08ec043] Mahabusaheb.P[4jn09ec040]
NCP meeting Jan 27-28, 2003, Brussels Colette Maloney Interfaces, Knowledge and Content technologies, Applications & Information Market DG INFSO Multimodal.
TOUCHLESS TOUCHSCREEN USER INTERFACE
Big Data – Lendület kutatócsoport Andras Benczur Insitute for Computer Science and Control Hungarian Academy of Sciences
Nickolas McCarley University of Alabama Abstract Robotic Navigation through Gesture Based Control (RNGBC) assists people who may not be able to operate.
Development of a Flex Sensor Glove
FINGERSPELLING THE SIGN LANGUAGE TOOL
REAL-TIME DETECTOR FOR UNUSUAL BEHAVIOR
Development of a Flex Sensor Glove
Development of a Flex Sensor Glove
CHAITANYA INSTITUTE OF SCIENCE AND TECHNOLOGY
Sliding Puzzle Project
Wala’ Hamad Khayrieh Homran

GESTURE RECOGNITION TECHNOLOGY
Toward an Accessible Community
Thrust IC: Action Selection in Joint-Human-Robot Teams
TOUCHLESS TOUCHSCREEN USER INTERFACE
Development of a Flex Sensor Glove
Ubi-Glove: A WISP-based Multi-purpose Ubiquitous Computing Tool
Development of a Flex Sensor Glove
Development of a Flex Sensor Glove
Automatic Translation
Development of a Flex Sensor Glove
Development of a Flex Sensor Glove
Plankton Classification VIDI: Sign Recognition HANDWRITING RECOGNITION
Development of a Flex Sensor Glove
The Accessible Webinar
Founded Meuro/year funding 12 FP6 and 20 FP5 IST projects
SENSOR BASED CONTROL OF AUTONOMOUS ROBOTS
Presentation transcript:

MTA SZTAKI An assistive interpreter tool using glove-based hand gesture recognition Department of Distributed Systems Péter Mátételki Máté Pataki Sándor Turbucz László Kovács My name is Peter Mátételki and let me present you our assistive interpreter tool that uses glove-based hand gesture recognition to translate sign language to speech.

pioneers of web technology in Hungary main fields MTA SZTAKI DSD MTA founded in 1827 SZTAKI founded in 1968 DSD founded in 1994 pioneers of web technology in Hungary main fields collaboration systems digital libraries web accessibility IoT crisis management plagiarism detection I am a research associate at the Hungarian Academy of Sciences, (MTA stands for it in Hungarian) Research Institute for Computer Science and Control, (abbreviated here as SZTAKI) Department of Distributed Systems.

Helen Keller Let me start with a quote from Helen Keller, the first deafblind person who earned a bachelor of arts degree, and who was also an activist supporting education for disabled people. Referring to her condition, she said: “Blindness separates us from things, but deafness separates us from people.” This very sentence leads us to the main problem of the deaf disabled group, and to the focus of our development. „Blindness separates us from things, but deafness separates us from people.”

hearing and speech impaired people Challenges Target groups hearing and speech impaired people deaf people (native language: sign language) Social problems problematic social integration isolated group To overcome the language barriers human sign language interpreter or: a good assistive tool Unlike for example the phisically disabled, hearing imparied people can not communicate with us. They speak another language, namely the sign language. For those who are deaf from birth, sign language is their native language. So when we try to speak with a deaf person, it’s the same situation as if we were trying to speak to a foreigner: we won’t understand each other because of the language barriers. It’s also a similar situation when you try to talk to your friend on a sidewalk of a busy road. You can’t hear each other so can not communicate, it’s very frustrating, right? This leads to serious social problems of the deaf, as they can not really integrate into the society, they form an isolated group. Today, their only chance to speak to the nondisabled is by being helped by a sign language interpreter. What our project aims is to give them a new assistive tool to let them communicate without any human assistance.

Silent Speech Translation (interACT) First, let’s see some other experiments in this field. The interACT project senses facial muscle movements with sensors attached to the skin and assigns text by matching the mimics to pre-recorded samples This solves the problem but probably most of you don’t want to walk on the street with these sensors glued to your face. Silent Speech Translation (interACT)

Eyes-Free Keypad Input Here you can see gesture controled keyboard. It is very simple: wherever you tap the display you get a 5. If you drag upwards you get a 2, if you drag downwards you get an 8 and so on. This is a great solution for numeric input but won’t work with a full keyboard to type letters. Eyes-Free Keypad Input

Keyboard Glove – University of Alabama The Keyboard Glove is made of a glove and micro-switches stitched under the glove. You can type by pushing the switches with your thumb. Here my problem is that this solution needs quite a bit of learning and I suspect that after a few hours of use your hand muscles will - most probably – be stiff, in pain. I think we need a more intuitive solution. Keyboard Glove – University of Alabama

is controlled by sign language (intuitive) talks for the deaf Talking Hands Our assistive tool is controlled by sign language (intuitive) talks for the deaf everyday use Features sign language to speech real time seamless Having seen the above projects our conclusion is that a good assistive tool for the deaf should be suitable for everyday use and follow their existing communication behavior. This means, that it should be controlled by sign language. So we suggest a solution for a real-time sign-language interpreter that translates the gestures to text and speech, realized in a way that can be used in everyday life. I hope that by now you became interested in our solution, so here it is:

TalkingHands We call it TalkingHands. Here you can see photos of our prototype showing sign language gestures. The solutions consists of a glove and a software component doing all the calculations as a mobile application.

custom gestures and text motion capture with the glove How does it work? sign language custom gestures and text motion capture with the glove gesture descriptor stream signal processing language processing text to speech Here is an overview of how it works: users can enter letters by showing hand gestures of the international fingerspelling alphabet AND can enter any text using custom gestures The glove captures the hand states transforms the hand state into gesture descriptors transmits the descriptors to the user’s mobile device The mobile application processes the gesture descriptors creates understandable text and reads it out loud On the following slides let me show you the glove and the processing algorithms with some more details.

Glove: gesture capturing 9DOF sensors signal fusion absolute position relative angles virtual hand gesture descriptors 30 Hagdil descriptor/sec Bluetooth For gesture capturing we use 3-axis accelerometers, gyroscopes and magnetometers on each major bones. We placed 2 sensors on each finger, 1 on the back of the hand and 1 on the wrist, a total of 12 sensors. We calculate the absolute position and the relative angles of the sensors, this results in a virtual hand. The virtual Hand is transformed to a gesture descriptor. For this we came up with our custom hand gesture descriptor language, that is called Hagdil. Each second, 30 Hagdil descriptors are transmitted to the mobile application. http://upload.wikimedia.org/wikipedia/commons/8/8c/Skeletal-hand_.jpg

Mobile application: processing Segmentation algorithm simple similarity repetition unkown descriptors sliding window kinetics based algorithm Context-sensitive auto-correction modified Levenshtein n-gram (1- & 2-grams) confusion matrix descriptors raw text text The Hagdil gesture descriptors are the input for the app. We need to pick the best ones from the stream, that’s the duty of the segmentation algorithm. We experimented with many approaches and found that the sliding window and the kinetics based algorithms produced the best results. The sliding window calculates an average in a sliding window. The kinetics based algorithm works by detecting the speed of the hand movements. We call the result of the segmentation raw text, as this text usually contains errors and typos. We tried existing spell-checkers to correct the raw text, but because of the different error characteristics they were not capable to do so. This is because the glove and segmentation errors are very different from the typos when typing on a keyboard. So we built a custom context-sensitive text-correction algorithm to transform the raw text into understandable text. Evaluation of these algorithms can be found in detail in the paper. (What we found the most interesting about the above algorithms, is that although the numeric results (so the distance of the original text and the raw text) resulting from the segmented stream are very similar, the context-sensitive correction algorithm can perform much better on the output of the kinetics based algorithm. This is caused by the two algorithm’s different correction- and error characteristics, so in the prototype we picked the more expensive segmentation algorithm.) speech

Scenarios, situations Can you guess, where can our TalkingHands help the deaf people?

work banking shopping public services healthcare education free time … Scenarios, situations work banking shopping public services healthcare education free time … Here are some scenarios that are, in general, problematic for the deaf people today. TalkingHands can enable them to handle normal everyday situations. They simply put on the glove, the phone in the pocket and signing can begin. (Lip reading Most hearing disabled people can lipread very well) Work: TalkingHands could improve employment opportunities for hearing disabled people, enhancing their social integration It does not only substitute a human signer (sign lang. Interpreter) but has a very positive impact on the user life (as the glove is at hand, they are not dependant on others) Grant for education: 120 hours + special curricula of 60 hours per semester

TalkingHands as a product two-handed gesture recognition Future plans TalkingHands as a product two-handed gesture recognition capture motion dynamics research in assistive technologies and IoT robotics remote manipulation Evolve Improve Enable

The project The project is executed as a consortium by two partners.

Péter Mátételki MTA SZTAKI DSD matetelki@sztaki. hu http://dsd. sztaki In case you are interested and need further information on TalkingHands, please contact me on the above address. I would be happy to answer all inquiries. Thank you for your attention!

www.youtube.com/watch?v=NhTUeZ16ZTw

Dactyl fingerspelling alphabet

Hagdil gesture descriptor

Architecture