Mobile Device and Cloud Server based Intelligent Health Monitoring Systems Sub-track in audio - visual processing NAME: ZHAO Ding SID: 52208367 Supervisor:

Slides:



Advertisements
Similar presentations
Face Recognition. Introduction Why we are interested in face recognition? Why we are interested in face recognition? Passport control at terminals in.
Advertisements

Face Recognition Method of OpenCV
Automatic Lip- Synchronization Using Linear Prediction of Speech Christopher Kohnert SK Semwal University of Colorado, Colorado Springs.
VOICE CONVERSION METHODS FOR VOCAL TRACT AND PITCH CONTOUR MODIFICATION Oytun Türk Levent M. Arslan R&D Dept., SESTEK Inc., and EE Eng. Dept., Boğaziçi.
AUTOMATIC ORGANIZING AND FORMATTING FOR LECTURE NOTES SHIQING (LICIA) HE ADIVISOR: PROF.KRISTINA STRIEGNITZ SPRING 2014 STRUCTURING THE UNSTRUCTURED NOTE:
GMM-Based Multimodal Biometric Verification Yannis Stylianou Yannis Pantazis Felipe Calderero Pedro Larroy François Severin Sascha Schimke Rolando Bonal.
Virtual Dart: An Augmented Reality Game on Mobile Device Supervisor: Professor Michael R. Lyu Prepared by: Lai Chung Sum Siu Ho Tung.
Personal Memory Assistant Abstract Facial recognition and speaker verification systems have been widely used in the security field. In this area the systems.
Department of Computer Science and Engineering, CUHK 1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal Video Conference Archives Indexing System.
Fig. 2 – Test results Personal Memory Assistant Facial Recognition System The facial identification system is divided into the following two components:
USER VERIFICATION SYSTEM. Scope Web Interface RGB separation Pervasive.
Computer-Aided Diagnosis and Display Lab Department of Radiology, Chapel Hill UNC Julien Jomier, Erwann Rault, and Stephen R. Aylward Computer.
LYU0103 Speech Recognition Techniques for Digital Video Library Supervisor : Prof Michael R. Lyu Students: Gao Zheng Hong Lei Mo.
Face Recognition: An Introduction
Digital signal Processing Digital signal Processing ECI Semester /2004 Telecommunication and Internet Engineering, School of Engineering, South.
Real-Time Face Detection and Tracking Using Multiple Cameras RIT Computer Engineering Senior Design Project John RuppertJustin HnatowJared Holsopple This.
A PRESENTATION BY SHAMALEE DESHPANDE
Oral Defense by Sunny Tang 15 Aug 2003
CP1610: Introduction to Computer Components
Computer Science AND DOCTORS Jolena Co Truong- 6 th period.
Facial Recognition CSE 391 Kris Lord.
Database Construction for Speech to Lip-readable Animation Conversion Gyorgy Takacs, Attila Tihanyi, Tamas Bardi, Gergo Feldhoffer, Balint Srancsik Peter.
Artificial Intelligence
Wen-Chyi Lin CS2310 Software Engineering.  “Never express yourself more clearly than you are able to think” by Niels Bohr. However, there are times and.
An FPGA implementation of real-time QRS detection H.K.Chatterjee Dept. of ECE Camellia School of Engineering & Technology Kolkata India R.Gupta, J.N.Bera,
Digital Sound and Video Chapter 10, Exploring the Digital Domain.
Kinect Player Gender Recognition from Speech Analysis
ACCURATE TELEMONITORING OF PARKINSON’S DISEASE SYMPTOM SEVERITY USING SPEECH SIGNALS Schematic representation of the UPDRS estimation process Athanasios.
Project 1 : Eigen-Faces Applied to Speech Style Classification Brad Keserich, Senior, Computer Engineering College of Engineering and Applied Science;
EWatchdog: An Electronic Watchdog for Unobtrusive Emotion Detection based on Usage Analysis Rayhan Shikder Department.
Automatic Pitch Tracking September 18, 2014 The Digitization of Pitch The blue line represents the fundamental frequency (F0) of the speaker’s voice.
Multimedia Databases (MMDB)
© 2003 East Collaborative e ast COLLABORATIVE ® eC SoftwareProducts TrackeCHealth.
Contactforum: Digitale bibliotheken voor muziek. 3/6/2005 Real music libraries in the virtual future: for an integrated view of music and music information.
August 12, 2004IAML - IASA 2004 Congress, Olso1 Music Information Retrieval, or how to search for (and maybe find) music and do away with incipits Michael.
KYLE PATTERSON Automatic Age Estimation and Interactive Museum Exhibits Advisors: Prof. Cass and Prof. Lawson.
Supervised by Prof. LYU, Rung Tsong Michael Department of Computer Science & Engineering The Chinese University of Hong Kong Prepared by: Chan Pik Wah,
Department of Computer Science and Engineering, CUHK 1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal Video Conference Archives Indexing System.
Face Recognition System By Arthur. Introduction  A facial recognition system is a computer application for automatically identifying or verifying a person.
Multimodal Information Analysis for Emotion Recognition
N n Debanga Raj Neog, Anurag Ranjan, João L. Cardoso, Dinesh K. Pai Sensorimotor Systems Lab, Department of Computer Science The University of British.
Real-Time Cyber Physical Systems Application on MobilityFirst Winlab Summer Internship 2015 Karthikeyan Ganesan, Wuyang Zhang, Zihong Zheng.
MULTIMEDIA INPUT / OUTPUT TECHNOLOGIES INTRODUCTION 6/1/ A.Aruna, Assistant Professor, Faculty of Information Technology.
Face Recognition: An Introduction
MACHINE VISION Machine Vision System Components ENT 273 Ms. HEMA C.R. Lecture 1.
Audio processing methods on marine mammal vocalizations Xanadu Halkias Laboratory for the Recognition and Organization of Speech and Audio
Facial Recognition Justin Kwong Megan Thompson Raymundo Vazquez-lugo.
Student Name: Honghao Chen Supervisor: Dr Jimmy Li Co-Supervisor: Dr Sherry Randhawa.
Counting How Many Words You Read
Performance Comparison of Speaker and Emotion Recognition
Computer Basics SystemsViruses Alternative Input Speech.
Automated Fingertip Detection
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Virtual Image Peephole By Kyle Patience Supervisor: Reg Dodds Co Supervisor: Mehrdad Ghaziasgar.
Augmented Reality and 3D modelling Done by Stafford Joemat Supervised by Mr James Connan.
1 What is Multimedia? Multimedia can have a many definitions Multimedia means that computer information can be represented through media types: – Text.
Frank Bergschneider February 21, 2014 Presented to National Instruments.
Emotion-aware recommender systems Master’s thesis SupervisorMentorStudent Katrien VerbertKarsten SeippJeroen Reinenbergh.
Portable Camera-Based Assistive Text and Product Label Reading From Hand-Held Objects for Blind Persons.
Face Detection 蔡宇軒.
Contents Introduction Requirements Design Technology Working Interaction.
AHED Automatic Human Emotion Detection
Submitted by: Ala Berawi Sujod Makhlof Samah Hanani Supervisor:
Recognition: Face Recognition
Interactive Visual System
Multimodal Caricatural Mirror
Project #2 Multimodal Caricatural Mirror Intermediate report
AHED Automatic Human Emotion Detection
AHED Automatic Human Emotion Detection
Elmar Nöth, Andreas Maier, Michael Stürmer, Maria Schuster Towards Multimodal Evaluation of Speech Pathologies Friday, 13 September 2019.
Presentation transcript:

Mobile Device and Cloud Server based Intelligent Health Monitoring Systems Sub-track in audio - visual processing NAME: ZHAO Ding SID: Supervisor: Prof YAN, Hong Assessor: Dr CHAN, Rosa H M

Objectives Develop an Android App:  To display the user’s talking speech pitch in the run time. To generate the pitch contour and pitch range analysis.  To measure the user’s heart rate using the built-in camera.  To recognize the user’s emotion status based on captured facial image and recorded daily for long-term monitoring.

Motivations  Fast life pace. Work stress. Inconvenient to visit hospital.  Chronic diseases and mental health problems. Essential to keep a record of daily emotion status.

Motivations Smartphones:  indispensible part of modern life.  Possible for health condition monitoring.

Work Done  Voice Disorder Checker  Heart Rate Monitor  Emotion Tracker

Work Done  Voice Disorder Checker  Heart Rate Monitor  Emotion Tracker

Voice Disorder Checker  Background Clinicians & subjective rating. Time-consuming. Special instrument or complex software. [1]

Voice Disorder Checker Record, sample and digitalize Pitch calculation and display sampling rate = Hz, encoding format = PCM 16 bit Feature extraction Timeframe: 46ms Pitch detection algorithms Alert for abnormal feature

Voice Disorder Checker  Pitch Detection Algorithms Direct Fast Fourier Transform Harmonic Product Spectrum [2] Cepstrum Analysis [3]

Voice Disorder Checker  Cepstrum Analysis Cepstrum of particular speech segment High-Key voice Low-Key voice Pitch contour over time (do re mi fa so la si do)

Voice Disorder Checker  Checking Results:[5] Abnormal FeaturesRelated Voice Disorders Unmatched pitch contour shape Dysprosody Reduced pitch range Vocal fold nodule, Vocal Hemorrhage Excessively high or low pitch Bogart–Bacall syndrome, Muscle Tension Dysphonia

Work Done  Voice Disorder Checker  Heart Rate Monitor  Emotion Tracker

Heart Rate Monitor  Background

Heart Rate Monitor Video record Heartbeat ++ Red pixel value > Avg value Heart Rate deduction Average red pixel intensity calculation Use PreviewCallback to grab the latest image Collect data in 10 sec chunk

 Image color intensity calculation YUV420SP != ARGB Heart Rate Monitor Y = luminance U and V = chrominance

Work Done  Voice Disorder Checker  Heart Rate Monitor  Emotion Tracker

Emotion Tracker  Background Static Approach FisherFace Model EigenFace Model [6] Active Appearance Model [7] Dynamic Approach FACS intensity tracking [8]

Emotion Tracker  Background Static Approach FisherFace Model EigenFace Model [6] Active Appearance Model [7] Dynamic Approach FACS intensity tracking [8]

Emotion Tracker Facial image capture Feed to EigenFace model trained Classification result recorded Long term monitoring report Model trained from JAFFE database

Emotion Tracker  EigenFace model Principal Component Analysis Training images from JAFFE database: Store training data in xml file Average Eigen Image Training images eigenfaces

Emotion Tracker  EigenFace model Load training data and test image Run the find nearest neighbor algorithm

Conclusions  VoiceDisorderChecker: Real-time speech pitch tracking.  HeartRateMonitor: Heartbeat counting. Red pixel intensity variation of index fingertip image, representative of blood pulse rhythm.  EmotionTracker: Static facial image expression recognition.

Work to be Done  Refine the pitch detection algorithm.  Evaluate the performance of EmotionTracker using figherface model.  More emotion categories when training eigenface model  Better design for App user interface  Release as beta version  Deploy the App to Google Cloud Platform

References [1] Koichi OMORI, “Diagnosis of Voice Disorders,” JMAJ, Vol. 54, No. 4, pp. 248–253, [2] TCH DETECTION METHODS REVIEW [Online]. Available: [3] A. Michael Noll, “Cepstrum Pitch Determination,” Journal of the Acoustical Society of America, Vol. 41, No. 2, (February 1967), pp [4] Alan V. Oppenheim and Ronald W. Schafer, Discrete-Time Signal Processing, Prentice Hall, [5] Deirdre D. Michael. (2012, Dec 1). Types of Voice Disorders. [Online]. Available:

References [6] Gender Classification with OpenCV. [Online]. Available at utorial/facerec_gender_classification.html#fisherfaces-for- gender-classification utorial/facerec_gender_classification.html#fisherfaces-for- gender-classification [7] Timothy F. Cootes, Gareth J. Edwards, and Christopher J. Taylor. “Active Appearance Models.” IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 23, NO. 6, JUNE [8] Maja Pantic, Student Member, IEEE, and Leon J.M. Rothkrantz. “Automatic Analysis of Facial Expressions: The State of the Art.” IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 22, NO. 12, DECEMBER 2000.

Q & A