Jacob Zurasky ECE5525 Fall 2010.  Goals ◦ Determine if the principles of speech processing relate to snoring sounds. ◦ Use homomorphic filtering techniques.

Slides:



Advertisements
Similar presentations
CS335 Principles of Multimedia Systems Audio Hao Jiang Computer Science Department Boston College Oct. 11, 2007.
Advertisements

Speech Processing AEGIS RET All-Hands Meeting University of Central Florida July 20, 2012 Applications of Images and Signals in High Schools.
Analysis and Digital Implementation of the Talk Box Effect Yuan Chen Advisor: Professor Paul Cuff.
Liner Predictive Pitch Synchronization Voiced speech detection, analysis and synthesis Jim Bryan Florida Institute of Technology ECE5525 Final Project.
Speech Processing AEGIS RET All-Hands Meeting University of Central Florida July 20, 2012 Applications of Images and Signals in High Schools.
Overview of Real-Time Pitch Tracking Approaches Music information retrieval seminar McGill University Francois Thibault.
Speech Recognition Chapter 3
Speech Sound Production: Recognition Using Recurrent Neural Networks Abstract: In this paper I present a study of speech sound production and methods for.
PH 105 Dr. Cecilia Vogel Lecture 14. OUTLINE  consonants  vowels  vocal folds as sound source  formants  speech spectrograms  singing.
Speech and Audio Processing and Recognition
SOME SIMPLE MANIPULATIONS OF SOUND USING DIGITAL SIGNAL PROCESSING Richard M. Stern demo August 31, 2004 Department of Electrical and Computer.
A STUDY ON SPEECH RECOGNITION USING DYNAMIC TIME WARPING CS 525 : Project Presentation PALDEN LAMA and MOUNIKA NAMBURU.
Pole Zero Speech Models Speech is nonstationary. It can approximately be considered stationary over short intervals (20-40 ms). Over thisinterval the source.
A STUDY ON SPEECH RECOGNITION USING DYNAMIC TIME WARPING CS 525 : Project Presentation PALDEN LAMA and MOUNIKA NAMBURU.
A PRESENTATION BY SHAMALEE DESHPANDE
Representing Acoustic Information
Digital Sound and Video Chapter 10, Exploring the Digital Domain.
DIGITAL VOICE NETWORKS ECE 421E Tuesday, October 02, 2012.
Kinect Player Gender Recognition from Speech Analysis
1 CS 551/651: Structure of Spoken Language Lecture 8: Mathematical Descriptions of the Speech Signal John-Paul Hosom Fall 2008.
Dual-Channel FFT Analysis: A Presentation Prepared for Syn-Aud-Con: Test and Measurement Seminars Louisville, KY Aug , 2002.
Speech Coding Using LPC. What is Speech Coding  Speech coding is the procedure of transforming speech signal into more compact form for Transmission.
Chapter 16 Speech Synthesis Algorithms 16.1 Synthesis based on LPC 16.2 Synthesis based on formants 16.3 Synthesis based on homomorphic processing 16.4.
Implementing a Speech Recognition System on a GPU using CUDA
93 SOURCE TIME SERIES VOCAL TRACT TRANSFER FUNCTION VOICE TIME SERIES VOICE SPECTRUM SOURCE TIME SERIES VOCAL TRACT TRANSFER FUNCTION VOICE TIME SERIES.
Jacob Zurasky ECE5526 – Spring 2011
Chapter 5: Speech Recognition An example of a speech recognition system Speech recognition techniques Ch5., v.5b1.
Speech Signal Representations I Seminar Speech Recognition 2002 F.R. Verhage.
Speaker Recognition by Habib ur Rehman Abdul Basit CENTER FOR ADVANCED STUDIES IN ENGINERING Digital Signal Processing ( Term Project )
Speech Recognition Feature Extraction. Speech recognition simplified block diagram Speech Capture Speech Capture Feature Extraction Feature Extraction.
ECE 5525 Osama Saraireh Fall 2005 Dr. Veton Kepuska
VOCODERS. Vocoders Speech Coding Systems Implemented in the transmitter for analysis of the voice signal Complex than waveform coders High economy in.
(Extremely) Simplified Model of Speech Production
Nick Kwolek David Duemeler Martin PendergastStephen Edwards.
A STUDY ON SPEECH RECOGNITION USING DYNAMIC TIME WARPING CS 525 : Project Presentation PALDEN LAMA and MOUNIKA NAMBURU.
EEL 6586: AUTOMATIC SPEECH PROCESSING Speech Features Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida February 27,
Vocal Tract & Lip Shape Estimation By MS Shah & Vikash Sethia Supervisor: Prof. PC Pandey EE Dept, IIT Bombay AIM-2003, EE Dept, IIT Bombay, 27 th June,
Speech Processing Homomorphic Signal Processing. 5 February 2016Veton Këpuska2 Outline  Principles of Homomorphic Signal Processing  Details of Homomorphic.
ECE 477 Final Presentation Team 13  Spring 2012 Martin Pendergast, Stephen Edwards, Nick Kwolek, David Duelmler.
By Sarita Jondhale 1 Signal preprocessor: “conditions” the speech signal s(n) to new form which is more suitable for the analysis Postprocessor: operate.
The Discrete Fourier Transform
Speech Processing Using HTK Trevor Bowden 12/08/2008.
Noise Reduction in Speech Recognition Professor:Jian-Jiun Ding Student: Yung Chang 2011/05/06.
Time Compression/Expansion Independent of Pitch. Listening Dies Irae from Requiem, by Michel Chion (1973)
Copyright ©2010, ©1999, ©1989 by Pearson Education, Inc. All rights reserved. Discrete-Time Signal Processing, Third Edition Alan V. Oppenheim Ronald W.
FREQUENCY (HZ) LOG10 MAGNITUDE Figure 2.30a. PSD of original voice
EEL 6586: AUTOMATIC SPEECH PROCESSING Speech Features Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida February 20,
Speech Recognition with Matlab ® Neil E. Cotter ECE Department UNIVERSITY OF UTAH
Phase Vocoder Marlon Smith. Split into amplitude and phase %do the windowing, stored in seperate %sections for FFT for m=0:1:(loopsentinel-(win/Ra)) for.
Speech Processing Dr. Veton Këpuska, FIT Jacob Zurasky, FIT.
Speech Processing AEGIS RET All-Hands Meeting
Ch. 5: Speech Recognition
Chapter 5 Homomorphic Processing(1)
Speech Processing AEGIS RET All-Hands Meeting
Cepstrum and MFCC Cepstrum MFCC Speech processing.
Linear Prediction.
Sharat.S.Chikkerur S.Anand Mantravadi Rajeev.K.Srinivasan
Digital Systems: Hardware Organization and Design
Linear Predictive Coding Methods
Mechanische trillingen
ECE 477 DESIGN REVIEW PART 2 Team 14
ESTIMATED INVERSE SYSTEM
Richard M. Stern demo January 12, 2009
Digital Systems: Hardware Organization and Design
Mark Hasegawa-Johnson 10/2/2018
Homomorphic Speech Processing
Speech Processing Final Project
EXP file structure.
Keyword Spotting Dynamic Time Warping
Music Signal Processing
Presentation transcript:

Jacob Zurasky ECE5525 Fall 2010

 Goals ◦ Determine if the principles of speech processing relate to snoring sounds. ◦ Use homomorphic filtering techniques to analyze snoring for pitch and also vocal tract response. ◦ Develop a method to distinguish a simple snore from a sleep apnea event.

 Store amplitude and frequency spectrum data to SD card  Interface to Sleep Lab polysomnogram equipment

 Top Figure is the frequency spectrum (0-2kHz)  Bottom figure is the snore amplitude

 Assume: s[n] = h[n] * p[n]  FFT -> log( ) -> IFFT, yields the cepstrum  Separate by low quefrency liftering  FFT -> exp( ) -> IFFT, vocal tract response

 Assume: s[n] = h[n] * p[n] (palatal flow)  Use sliding hamming window, 50% overlap  Analyze different sounds clips for differences in h[n] and p[n] between normal snoring and an apnea event.

 p[n], ‘Voicing’, of the sleep apnea files has a much larger magnitude in the cepstral domain.  Vocal tract response during a simple snore is more stable than during an apnea.  Vocal tract response is slower changing during a simple snore.

 Redesign the SRD to incorporate the functions of the MATLAB code.  Faster processor, floating point architecture  Continue research to develop a method for in home screening of sleep apnea.