Real-time Implementation of Multi-band Frequency Compression for Listeners with Moderate Sensorineural Impairment [Ref.: N. Tiwari, P. C. Pandey, P. N.

Slides:



Advertisements
Similar presentations
Time-Frequency Analysis Analyzing sounds as a sequence of frames
Advertisements

August 2004Multirate DSP (Part 2/2)1 Multirate DSP Digital Filter Banks Filter Banks and Subband Processing Applications and Advantages Perfect Reconstruction.
Advanced Speech Enhancement in Noisy Environments
Guitar Effects Processor Using DSP
Feedback Cancellation in Public Address systems using notch filters Shailesh Kulkarni Vaibhav Mathur Digital Signal Processing-Advanced Topics.
Page 0 of 34 MBE Vocoder. Page 1 of 34 Outline Introduction to vocoders MBE vocoder –MBE Parameters –Parameter estimation –Analysis and synthesis algorithm.
EE Dept., IIT Bombay Workshop “AICTE Sponsored Faculty Development Programme on Signal Processing and Applications", Dept. of Electrical.
VOICE CONVERSION METHODS FOR VOCAL TRACT AND PITCH CONTOUR MODIFICATION Oytun Türk Levent M. Arslan R&D Dept., SESTEK Inc., and EE Eng. Dept., Boğaziçi.
1 Audio Compression Techniques MUMT 611, January 2005 Assignment 2 Paul Kolesnik.
FFT-based filtering and the Short-Time Fourier Transform (STFT) R.C. Maher ECEN4002/5002 DSP Laboratory Spring 2003.
Speech Enhancement Based on a Combination of Spectral Subtraction and MMSE Log-STSA Estimator in Wavelet Domain LATSI laboratory, Department of Electronic,
MPEG Audio Compression by V. Loumos. Introduction Motion Picture Experts Group (MPEG) International Standards Organization (ISO) First High Fidelity Audio.
Voice Transformations Challenges: Signal processing techniques have advanced faster than our understanding of the physics Examples: – Rate of articulation.
Warped Linear Prediction Concept: Warp the spectrum to emulate human perception; then perform linear prediction on the result Approaches to warp the spectrum:
A Full Frequency Masking Vocoder for Legal Eavesdropping Conversation Recording R. F. B. Sotero Filho, H. M. de Oliveira (qPGOM), R. Campello de Souza.
Representing Acoustic Information
Normalization of the Speech Modulation Spectra for Robust Speech Recognition Xiong Xiao, Eng Siong Chng, and Haizhou Li Wen-Yi Chu Department of Computer.
Ni.com Data Analysis: Time and Frequency Domain. ni.com Typical Data Acquisition System.
GCT731 Fall 2014 Topics in Music Technology - Music Information Retrieval Overview of MIR Systems Audio and Music Representations (Part 1) 1.
Figures for Chapter 6 Compression
EE Dept., IIT Bombay NCC2014 Kanpur, 28 Feb.- 2 Mar. 2014, Paper No (Session III, Sat., 1 st Mar., 1020 – 1200) A Sliding-band.
Speech Enhancement Using Noise Estimation Based on
1 SPEECH PROCESSING FOR BINAURAL HEARING AIDS Dr P. C. Pandey EE Dept., IIT Bombay Feb’03.
Real-time Enhancement of Noisy Speech Using Spectral Subtraction
Second International Conference on Intelligent Interactive Technologies and Multimedia (IITM 2013), March 2013, Allahabad, India 09 March 2013 Speech.
Audio Compression Usha Sree CMSC 691M 10/12/04. Motivation Efficient Storage Streaming Interactive Multimedia Applications.
IIT Bombay Dr. Prem C. Pandey Dr. Pandey is a Professor in Electrical Engineering at IIT Bombay. He is currently also the Associate.
A Tutorial on MPEG/Audio Compression Davis Pan, IEEE Multimedia Journal, Summer 1995 Presented by: Randeep Singh Gakhal CMPT 820, Spring 2004.
♠♠♠♠ 1Intro 2.Loudness 3.Method. 4.Results 5.Concl. ♦♦ ◄◄ ► ► 1/161Intro 2.Loudness 3.Method. 4.Results 5.Concl. ♦♦ ◄ ► IIT Bombay ICA 2010 : 20th Int.
Speech Enhancement Using Spectral Subtraction
Page 0 of 23 MELP Vocoders Nima Moghadam SN#: Saeed Nari SN#: Supervisor Dr. Saameti April 2005 Sharif University of Technology.
Chapter 16 Speech Synthesis Algorithms 16.1 Synthesis based on LPC 16.2 Synthesis based on formants 16.3 Synthesis based on homomorphic processing 16.4.
EE Dept., IIT Bombay Indicon2013, Mumbai, Dec. 2013, Paper No. 524 (Track 4.1,
SPEECH CODING Maryam Zebarjad Alessandro Chiumento.
1/18 1.Intro 2. Implementation 3. Results 4. Con.
By Sarita Jondhale1 Signal Processing And Analysis Methods For Speech Recognition.
EE Dept., IIT Bombay NCC 2013, Delhi, Feb. 2013, Paper 3.2_2_ ( Sat.16 th, 1135 – 1320, 3.2_2) Speech Enhancement.
EE Dept., IIT Bombay NCC 2015, Mumbai, 27 Feb.- 1 Mar. 2015, Paper No (28 th Feb., Sat., Session SI, 10:05 – 11:15, Paper.
♠ 1.Intro 2. List. tests 3. Results 4 Concl.♠♠ 1.Intro 2. List. tests 3. Results 4 Concl. ♥♥ ◄◄ ► ► 1/17♥♥◄ ► IIT Bombay ICA 2010 : 20th Int. Congress.
Nico De Clercq Pieter Gijsenbergh.  Problem  Solutions  Single-channel approach  Multichannel approach  Our assignment Overview.
Speech Signal Representations I Seminar Speech Recognition 2002 F.R. Verhage.
Authors: Sriram Ganapathy, Samuel Thomas, and Hynek Hermansky Temporal envelope compensation for robust phoneme recognition using modulation spectrum.
Image Processing Architecture, © 2001, 2002, 2003 Oleh TretiakPage 1 ECE-C490 Image Processing Architecture MP-3 Compression Course Review Oleh Tretiak.
Gammachirp Auditory Filter
EE Dept., IIT Bombay IEEE Workshop on Intelligent Computing, IIIT Allahabad, Oct Signal processing for improving speech.
ECE 5525 Osama Saraireh Fall 2005 Dr. Veton Kepuska
VOCODERS. Vocoders Speech Coding Systems Implemented in the transmitter for analysis of the voice signal Complex than waveform coders High economy in.
ITU-T G.729 EE8873 Rungsun Munkong March 22, 2004.
P. N. Kulkarni, P. C. Pandey, and D. S. Jangamashetti / DSP 2009, Santorini, 5-7 July DSP 2009 (Santorini, Greece. 5-7 July 2009), Session: S4P,
IIT Bombay {pcpandey,   Intro. Proc. Schemes Evaluation Results Conclusion Intro. Proc. Schemes Evaluation Results Conclusion.
EE Dept., IIT Bombay P. C. Pandey, "Signal processing for persons with sensorineural hearing loss: Challenges and some solutions,”
Present document contains informations proprietary to France Telecom. Accepting this document means for its recipient he or she recognizes the confidential.
EE Dept., IIT Bombay Part B Sliding-band Dynamic Range Compression (N. Tiwari & P. C. Pandey, NCC 2014) P. C. Pandey, "Signal processing.
EE Dept., IIT Bombay Workshop “Radar and Sonar Signal Processing,” NSTL Visakhapatnam, Aug 2015 Coordinator: Ms. M. Vijaya.
1 Introduction1 Introduction 2 Noise red. tech 3 Spect. Subtr. 4. QBNE 5 Invest. QBNE 6 Conc., & future work2 Noise red. tech 3 Spect. Subtr.4. QBNE5 Invest.
IIT Bombay 17 th National Conference on Communications, Jan. 2011, Bangalore, India Sp Pr. 1, P3 1/21 Detection of Burst Onset Landmarks in Speech.
A. R. Jayan, P. C. Pandey, EE Dept., IIT Bombay 1 Abstract Perception of speech under adverse listening conditions may be improved by processing it to.
1 Introduction1 Introduction 2 Spectral subtraction 3 QBNE 4 Results 5 Conclusion, & future work2 Spectral subtraction 3 QBNE4 Results5 Conclusion, & future.
EE Dept., IIT Bombay CEP-cum-TEQUIP-KITE Course “Digital Signal Processing”, IIT Bombay, 2–6 November 2015, Course Coordinator:
Codec 2 ● open source speech codec ● low bit rate (2400 bit/s and below) ● applications include digital speech for HF and VHF radio ● fills gap in open.
Saketh Sharma, Nitya Tiwari, & Prem C. Pandey
COMPUTER NETWORKS and INTERNETS
ARTIFICIAL NEURAL NETWORKS
Single-channel Speech Enhancement for Real-time Applications
Embedded Systems Design
A Smartphone App-Based
Linear Predictive Coding Methods
Results from offline processing
MPEG-1 Overview of MPEG-1 Standard
Govt. Polytechnic Dhangar(Fatehabad)
Presentation transcript:

Real-time Implementation of Multi-band Frequency Compression for Listeners with Moderate Sensorineural Impairment [Ref.: N. Tiwari, P. C. Pandey, P. N. Kulkarni, Proc. Interspeech 2012, Portland, Oregon, 9-13 Sept 2012, Paper 689]

Outline 1.Introduction 2.Signal processing 3.Real-time implementation 4.Results 5.Conclusion 1.Introduction 2.Signal processing 3.Real-time implementation 4.Results 5.Conclusion 2/24

/24 Signal processing for reducing the effects of increased intra-speech spectral masking Binaural hearing (moderate bilateral loss) – Binaural dichotic presentation (Lunner et a l. 1993, Kulkarni et al. 2012) Monaural hearing – Spectral contrast enhancement (Yang et al. 2003) Errors in identifying peaks, increases dynamic range >> degraded speech perception – Multi-band frequency compression (Arai et al. 2004, Kulkarni et al. 2012) Multi-band frequency compression Presentation of speech energy in relatively narrow bands to avoid masking by adjacent spectral components, and without causing perceptible distortions. Signal processing for reducing the effects of increased intra-speech spectral masking Binaural hearing (moderate bilateral loss) – Binaural dichotic presentation (Lunner et a l. 1993, Kulkarni et al. 2012) Monaural hearing – Spectral contrast enhancement (Yang et al. 2003) Errors in identifying peaks, increases dynamic range >> degraded speech perception – Multi-band frequency compression (Arai et al. 2004, Kulkarni et al. 2012) Multi-band frequency compression Presentation of speech energy in relatively narrow bands to avoid masking by adjacent spectral components, and without causing perceptible distortions. 1. Introduction

Research objective Implementation of multi-band frequency compression for real-time operation for use in hearing aids Reduction in computational requirement for implementing on a DSP chip without speech quality degradation Low algorithmic and computational delay for real-time implementation on a DSP board with16-bit fixed-point processor Research objective Implementation of multi-band frequency compression for real-time operation for use in hearing aids Reduction in computational requirement for implementing on a DSP chip without speech quality degradation Low algorithmic and computational delay for real-time implementation on a DSP board with16-bit fixed-point processor 4/24

Signal Processing Signal processing for multi-band frequency compression Processing steps Segmentation and spectral analysis ● Spectral modification Resynthesis Multi-band frequency compression (Arai et al. 2004) Processing using auditory critical bandwidth Compressed magnitude & original phase spectrum >> Complex spectrum Multi-band frequency compression (Kulkarni et al. 2009, 2012) Investigations using different bandwidth, segmentation & frequency mapping Processing using complex spectrum Reduced computation and processing artifacts Signal processing for multi-band frequency compression Processing steps Segmentation and spectral analysis ● Spectral modification Resynthesis Multi-band frequency compression (Arai et al. 2004) Processing using auditory critical bandwidth Compressed magnitude & original phase spectrum >> Complex spectrum Multi-band frequency compression (Kulkarni et al. 2009, 2012) Investigations using different bandwidth, segmentation & frequency mapping Processing using complex spectrum Reduced computation and processing artifacts 5/24

Multi-band frequency compression (Kulkarni et al. 2009, 2012) Signal processing Segmentation using analysis window with 50 % overlap – fixed-frame (FF) segmentation : window length L = 20 ms – pitch-synchronous (PS) segmentation: L = 2 pitch periods / prev. L (voiced / unvoiced) Zero padding >> N -point frame Complex spectra by N -point DFT Analysis bands for compression – fixed bandwidth – one-third octave – auditory critical band (ACB) Mappings for spectral modification – sample-to-sample – spectral sample superposition – spectral segment Resynthesis by overlap-add method Multi-band frequency compression (Kulkarni et al. 2009, 2012) Signal processing Segmentation using analysis window with 50 % overlap – fixed-frame (FF) segmentation : window length L = 20 ms – pitch-synchronous (PS) segmentation: L = 2 pitch periods / prev. L (voiced / unvoiced) Zero padding >> N -point frame Complex spectra by N -point DFT Analysis bands for compression – fixed bandwidth – one-third octave – auditory critical band (ACB) Mappings for spectral modification – sample-to-sample – spectral sample superposition – spectral segment Resynthesis by overlap-add method 6/24 Frequency mapping with compression factor α = 0.6 as shown by Kulkarni et al. 2012

Details of signal processing Analysis band for compression 18 bands based on ACB Mapping for spectral modification Spectral segment mapping No irregular variations in spectrum Resynthesis Resynthesis by N -point IDFT Reconstruction by overlap-add method Details of signal processing Analysis band for compression 18 bands based on ACB Mapping for spectral modification Spectral segment mapping No irregular variations in spectrum Resynthesis Resynthesis by N -point IDFT Reconstruction by overlap-add method 7/24 Spectral Segment Mapping as shown by Kulkarni et al (a) Unprocessed /i/ BB noise (b) Processed Unprocessed and processed (α =0.6) spectra of vowel /i/ and broad band nois e

/24 Evaluation and results Modified Rhyme Test (MRT) Best results with PS segmentation, ACB, spectral segment mapping & compression factor of 0.6 Normal hearing subjects with loss simulated by additive noise – 17 % improvement in recognition scores – 0.88 s decrease in response time Subjects with moderate sensorineural loss – 16.5 % improvement in recognition scores – 0.89 s decrease in response time Evaluation and results Modified Rhyme Test (MRT) Best results with PS segmentation, ACB, spectral segment mapping & compression factor of 0.6 Normal hearing subjects with loss simulated by additive noise – 17 % improvement in recognition scores – 0.88 s decrease in response time Subjects with moderate sensorineural loss – 16.5 % improvement in recognition scores – 0.89 s decrease in response time

/24 Real-time implementation requirements Comparison of processing schemes with different segmentation FF processing: Fixed-frame segmentation Discontinuities masked by 50 % overlap-add Perceptible distortions in form of another superimposed pitch related to shift duration PS Processing: Pitch-synchronous segmentation Less perceptible distortions Pitch estimation (large computational requirement) Not suitable for music Modified segmentation scheme for real-time implementation Lower perceptual distortions Lower computational requirement Real-time implementation requirements Comparison of processing schemes with different segmentation FF processing: Fixed-frame segmentation Discontinuities masked by 50 % overlap-add Perceptible distortions in form of another superimposed pitch related to shift duration PS Processing: Pitch-synchronous segmentation Less perceptible distortions Pitch estimation (large computational requirement) Not suitable for music Modified segmentation scheme for real-time implementation Lower perceptual distortions Lower computational requirement

Modified multi-band frequency compression (LSEE processing) Griffin and Lim’s method of signal estimation from modified STFT Minimize mean square error between estimated and modified STFT Multiplication by analysis window before overlap-add Window requirement For partial overlap only a few windows meet the requirement Fixed window length L and shift S, with S = L / 4 (75% overlap) Modified Hamming window where p = 0.54 and q = Modified multi-band frequency compression (LSEE processing) Griffin and Lim’s method of signal estimation from modified STFT Minimize mean square error between estimated and modified STFT Multiplication by analysis window before overlap-add Window requirement For partial overlap only a few windows meet the requirement Fixed window length L and shift S, with S = L / 4 (75% overlap) Modified Hamming window where p = 0.54 and q = /24

Comparison of analysis-synthesis techniques for multi-band frequency compression Pitch-synchronous implementation LSEE method based implementation Evaluation Informal listening Objective evaluation using PESQ measure (0 – 4.5) Test material Sentence “ Where were you a year ago?” from a male speaker, vowels /a/-/i/-/u/, music Results PESQ score (compression factor range 1 ― 0.6): 4.3 ― 3.7 (sentence/ vowels) Comparison of analysis-synthesis techniques for multi-band frequency compression Pitch-synchronous implementation LSEE method based implementation Evaluation Informal listening Objective evaluation using PESQ measure (0 – 4.5) Test material Sentence “ Where were you a year ago?” from a male speaker, vowels /a/-/i/-/u/, music Results PESQ score (compression factor range 1 ― 0.6): 4.3 ― 3.7 (sentence/ vowels) 11/24 Investigations using offline implementation a) b) c) d) Comparison of analysis-synthesis methods: (a) unprocessed, (b) FF, (c) PS, and (d) LSEE. Processing with spectral segment mapping, auditory critical bandwidth, α = 0.6.

Input Signal Fixed-frame MFCPitch-synchronousFixed-frame LSEE α = 1α = 0.8 α = 0.6 α = 1α = 0.8 α = 0.6 α = 1α = 0.8 α = 0.6 Vowels /a i u / “Where were you a year ago?” Music clip Conclusion LSEE processing suited for non-speech and audio applications 12/24 Processed outputs from off-line implementation

/24 3. Real-time Implementation 16-bit fixed point DSP: TI/TMS320C MB memory space : 320 KB on-chip RAM with 64 KB DARAM, 128 KB on-chip ROM Three 32-bit programmable timers, 4 DMA controllers each with 4 channels FFT hardware accelerator (8 to 1024-point FFT) Max. clock speed: 120 MHz DSP Board: eZdsp 4 MB on-board NOR flash for user program Codec TLV320AIC3204: stereo ADC & DAC, 16/20/24/32-bit quantization, 8 – 192 kHz sampling Development environment for C: TI's 'CCStudio, ver. 4.0‘

/24 Implementation details ADC & DAC quantization: 16-bit ● Sampling rate: 10 kHz FFT length N : 1024 / 512 Analysis-synthesis : LSEE-based fixed frame, 260-sample (26 ms) window, 75% overlap Mapping for spectral modification: spectral segment mapping with auditory critical bands Input samples, spectral values & processed output: 32-bits each with 16-bit real and imaginary parts Implementation details ADC & DAC quantization: 16-bit ● Sampling rate: 10 kHz FFT length N : 1024 / 512 Analysis-synthesis : LSEE-based fixed frame, 260-sample (26 ms) window, 75% overlap Mapping for spectral modification: spectral segment mapping with auditory critical bands Input samples, spectral values & processed output: 32-bits each with 16-bit real and imaginary parts Signal acquisition & processing

/24 Data transfer and buffering operations (S = L/4) DMA cyclic buffers 5 block input buffer 2 block output buffer (each of S samples) Four pointers current input block current output block just-filled input block write-to output block (Pointers incremented cyclically on DMA interrupt) Efficient realization of 75 % overlap & zero padding Total delay Algorithmic delay: L = 4 S samples (26 ms) Computational delay: L /4 = S samples (6.5 ms) Data transfer and buffering operations (S = L/4) DMA cyclic buffers 5 block input buffer 2 block output buffer (each of S samples) Four pointers current input block current output block just-filled input block write-to output block (Pointers incremented cyclically on DMA interrupt) Efficient realization of 75 % overlap & zero padding Total delay Algorithmic delay: L = 4 S samples (26 ms) Computational delay: L /4 = S samples (6.5 ms)

Comparison of proc. output from offline-implementation & DSP board Test material Sentence “ Where were you a year ago?”, vowels /a/-/i/-/u/, music Evaluation methods Informal listening ● PESQ measure (0 – 4.5) Results Lowest clock for satisfactory operation: 20 MHz / 12 MHz ( N = 1024/512) No perceptible change in output with N = 512, processing capacity used ≈ 1/10 of the capacity with highest clock (120 MHz) Informal listening: Processed output from DSP similar to corresponding output from offline implementation PESQ scores (compression factor range 0.6 ― 1) For sentence 2.5 ― 3.4 & for vowels 3.0 ― 3.4 Total processing delay: 35 ms Comparison of proc. output from offline-implementation & DSP board Test material Sentence “ Where were you a year ago?”, vowels /a/-/i/-/u/, music Evaluation methods Informal listening ● PESQ measure (0 – 4.5) Results Lowest clock for satisfactory operation: 20 MHz / 12 MHz ( N = 1024/512) No perceptible change in output with N = 512, processing capacity used ≈ 1/10 of the capacity with highest clock (120 MHz) Informal listening: Processed output from DSP similar to corresponding output from offline implementation PESQ scores (compression factor range 0.6 ― 1) For sentence 2.5 ― 3.4 & for vowels 3.0 ― 3.4 Total processing delay: 35 ms 16/24 4. Results

Input Signal Fixed-frame LSEE MFC α = 1α = 0.8α = 0.6 Vowels /a/-/i/-/u/ “Where were you a year ago?” Music clip 17/24 Processed outputs from DSP board

Multi-band frequency compression using LSEE processing -- Processed speech output perceptually similar to that from earlier established pitch- synchronous processing. -- Suitable for speech as well as non-speech audio signals. -- Well-suited for real-time processing due to use of fixed-frame segmentation. Implementation for real-time operation using 16-bit fixed-point processor TI/TMS320C5515: used-up processing capacity ≈ 1/10, delay = 35 ms Further work Combining multi-band frequency compression with other types of processing for use in hearing aids Frequency-selective gain Multi-band dynamic range compression (settable gain and compression ratios) Noise reduction CVR modification Multi-band frequency compression using LSEE processing -- Processed speech output perceptually similar to that from earlier established pitch- synchronous processing. -- Suitable for speech as well as non-speech audio signals. -- Well-suited for real-time processing due to use of fixed-frame segmentation. Implementation for real-time operation using 16-bit fixed-point processor TI/TMS320C5515: used-up processing capacity ≈ 1/10, delay = 35 ms Further work Combining multi-band frequency compression with other types of processing for use in hearing aids Frequency-selective gain Multi-band dynamic range compression (settable gain and compression ratios) Noise reduction CVR modification 18/24 5. Conclusion

THANK YOU

/ Abstract Widening of auditory filters in persons with sensorineural hearing impairment leads to increased spectral masking and degraded speech perception. Multi-band frequency compression of the complex spectral samples using pitch-synchronous processing has been reported to increase speech perception by persons with moderate sensorineural loss. It is shown that implementation of multi-band frequency compression using fixed-frame processing along with least-squares error based signal estimation reduces the processing delay and the speech output is perceptually similar to that from pitch-synchronous processing. The processing is implemented on a DSP board based on the 16-bit fixed-point processor TMS320C5515, and real-time operation is achieved using about one-tenth of its computing capacity.

/ References [1]H. Levitt, J. M. Pickett, and R. A. Houde (eds.), Senosry Aids for the Hearing Impaired. New York: IEEE Press, [2]B. C. J. Moore, An Introduction to the Psychology of Hearing, London, UK: Academic, 1997, pp 66–107. [3]J. M. Pickett, The Acoustics of Speech Communication: Fundamentals, Speech Perception Theory, and Technology. Boston, Mass.: Allyn Bacon, 1999, pp. 289–323. [4]H. Dillon, Hearing Aids. New York: Thieme Medical, [5]T. Lunner, S. Arlinger, and J. Hellgren, “8-channel digital filter bank for hearing aid use: preliminary results in monaural, diotic, and dichotic modes,” Scand. Audiol. Suppl., vol. 38, pp. 75–81, [6]P. N. Kulkarni, P. C. Pandey, and D. S. Jangamashetti, “Binaural dichotic presentation to reduce the effects of spectral masking in moderate bilateral sensorineural hearing loss”, Int. J. Audiol., vol. 51, no. 4, pp. 334–344, [7]T. Baer, B. C. J. Moore, and S. Gatehouse, “Spectral contrast enhancement of speech in noise for listeners with sensorineural hearing impairment: effects on intelligibility, quality, and response times”, Int. J. Rehab. Res., vol. 30, no. 1, pp. 49–72, [8]J. Yang, F. Luo, and A. Nehorai, “Spectral contrast enhancement: Algorithms and comparisons,” Speech Commun., vol. 39, no. 1–2, pp. 33–46, [9]T. Arai, K. Yasu, and N. Hodoshima, “Effective speech processing for various impaired listeners,” in Proc. 18th Int. Cong. Acoust., 2004, Kyoto, Japan, pp. 1389–1392. [10]K. Yasu, M. Hishitani, T. Arai, and Y. Murahara, “Critical-band based frequency compression for digital hearing aids,” Acoustical Science and Technology, vol. 25, no. 1, pp , [11]P. N. Kulkarni, P. C. Pandey, and D. S. Jangamashetti, “Multi-band frequency compression for reducing the effects of spectral masking,” Int. J. Speech Tech., vol. 10, no. 4, pp. 219–227, [12]P. N. Kulkarni, P. C. Pandey, and D. S. Jangamashetti, “Multi-band frequency compression for improving speech perception by listeners with moderate sensorineural hearing loss,” Speech Commun., vol. 54, no. 3 pp. 341–350, [13]E. Zwicker, “Subdivision of the audible frequency range into critical bands (Freqenzgruppen),” J. Acoust. Soc. Am., vol. 33, no. 2, pp. 248, [14]D. G. Childers and H. T. Hu, “Speech synthesis by glottal excited linear prediction”, J. Acoust. Soc. Am., vol. 96, no. 4, pp. 2026–2036, 1994.

/ [15]D. W. Griffin and J. S. Lim, “Signal estimation from modified short-time Fourier transform,” IEEE Trans. Acoustics, Speech, Signal Proc., vol. 32, no. 2, pp , [16]X. Zhu, G. T. Beauregard, and L. L. Wyse, “Real-time signal estimation from modified short-time Fourier transform magnitude spectra”, IEEE Trans. Audio, Speech, Language Process., vol. 15, no. 5, pp. 1645–1653, [17]Texas Instruments Inc., TMS320C5515 Fixed-Point Digital Signal Processor. 2011, [online] Available: lit/ds/symlink/tms320c5515.pdf. [18] Texas Instruments Inc., TLV320AIC3204 Ultra Low Power Stereo Audio Codec. 2008, [online] Available: ti.com/lit/ds/symlink/tlv320aic3204.pdf.