Digital Signal Processing of Speech for the Hearing Impaired

Similar documents
GSM Interference Cancellation For Forensic Audio

The University of Texas at Austin Dept. of Electrical and Computer Engineering Midterm #1

CHAPTER 2 FIR ARCHITECTURE FOR THE FILTER BANK OF SPEECH PROCESSOR

EECS 452 Midterm Exam Winter 2012

Part One. Efficient Digital Filters COPYRIGHTED MATERIAL

SCUBA-2. Low Pass Filtering

Problem Point Value Your score Topic 1 28 Filter Analysis 2 24 Filter Implementation 3 24 Filter Design 4 24 Potpourri Total 100

CG401 Advanced Signal Processing. Dr Stuart Lawson Room A330 Tel: January 2003

SIGMA-DELTA CONVERTER

DSP Laboratory (EELE 4110) Lab#10 Finite Impulse Response (FIR) Filters

Corso di DATI e SEGNALI BIOMEDICI 1. Carmelina Ruggiero Laboratorio MedInfo

Aparna Tiwari, Vandana Thakre, Karuna Markam Deptt. Of ECE,M.I.T.S. Gwalior, M.P, India

The University of Texas at Austin Dept. of Electrical and Computer Engineering Final Exam

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter

Keywords: Adaptive filtering, LMS algorithm, Noise cancellation, VHDL Design, Signal to noise ratio (SNR), Convergence Speed.

Adaptive Filters Application of Linear Prediction

Different Approaches of Spectral Subtraction Method for Speech Enhancement

BIOMEDICAL DIGITAL SIGNAL PROCESSING

EE 403: Digital Signal Processing

A Low-Power Broad-Bandwidth Noise Cancellation VLSI Circuit Design for In-Ear Headphones

Analysis on Extraction of Modulated Signal Using Adaptive Filtering Algorithms against Ambient Noises in Underwater Communication

ASN Filter Designer Professional/Lite Getting Started Guide

ECE438 - Laboratory 7a: Digital Filter Design (Week 1) By Prof. Charles Bouman and Prof. Mireille Boutin Fall 2015

Auditory modelling for speech processing in the perceptual domain

(i) Understanding of the characteristics of linear-phase finite impulse response (FIR) filters

AC : INTERACTIVE LEARNING DISCRETE TIME SIGNALS AND SYSTEMS WITH MATLAB AND TI DSK6713 DSP KIT

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm

Contents. Introduction 1 1 Suggested Reading 2 2 Equipment and Software Tools 2 3 Experiment 2

WARPED FILTER DESIGN FOR THE BODY MODELING AND SOUND SYNTHESIS OF STRING INSTRUMENTS

(i) Understanding the basic concepts of signal modeling, correlation, maximum likelihood estimation, least squares and iterative numerical methods

EE 422G - Signals and Systems Laboratory

Practical applications of digital filters

Discrete-Time Signal Processing (DTSP) v14

Speech Enhancement using Wiener filtering

Presented at the 108th Convention 2000 February Paris, France

FAST ADAPTIVE DETECTION OF SINUSOIDAL SIGNALS USING VARIABLE DIGITAL FILTERS AND ALL-PASS FILTERS

(i) Understanding of the characteristics of linear-phase finite impulse response (FIR) filters

An Overview of the Decimation process and its VLSI implementation

Speech Enhancement Using Spectral Flatness Measure Based Spectral Subtraction

MODIFIED DCT BASED SPEECH ENHANCEMENT IN VEHICULAR ENVIRONMENTS

Real-time adaptive filtering of dental drill noise using a digital signal processor

Flatten DAC frequency response EQUALIZING TECHNIQUES CAN COPE WITH THE NONFLAT FREQUENCY RESPONSE OF A DAC.

EE482: Digital Signal Processing Applications

ECE Digital Signal Processing

Development of Real-Time Adaptive Noise Canceller and Echo Canceller

EECS 452 Midterm Exam (solns) Fall 2012

Signals and Filtering

Understanding Digital Signal Processing

B.Tech III Year II Semester (R13) Regular & Supplementary Examinations May/June 2017 DIGITAL SIGNAL PROCESSING (Common to ECE and EIE)

DSP Based Corrections of Analog Components in Digital Receivers

IMPLEMENTATION OF DIGITAL FILTER ON FPGA FOR ECG SIGNAL PROCESSING

EFFECTS OF PHYSICAL CONFIGURATIONS ON ANC HEADPHONE PERFORMANCE

A Computational Efficient Method for Assuring Full Duplex Feeling in Hands-free Communication

Signals & Systems for Speech & Hearing. Week 6. Practical spectral analysis. Bandpass filters & filterbanks. Try this out on an old friend

CS3291: Digital Signal Processing

Design of FIR Filter for Efficient Utilization of Speech Signal Akanksha. Raj 1 Arshiyanaz. Khateeb 2 Fakrunnisa.Balaganur 3

Team proposals are due tomorrow at 6PM Homework 4 is due next thur. Proposal presentations are next mon in 1311EECS.

A Digital Signal Processor for Musicians and Audiophiles Published on Monday, 09 February :54

ELT Receiver Architectures and Signal Processing Fall Mandatory homework exercises

Lab 4 Digital Scope and Spectrum Analyzer

ELEC-C5230 Digitaalisen signaalinkäsittelyn perusteet

Active Noise Cancellation System Using DSP Prosessor

Lecture 3 Review of Signals and Systems: Part 2. EE4900/EE6720 Digital Communications

A Survey on Power Reduction Techniques in FIR Filter

EECS 452 Midterm Closed book part Winter 2013

Encoding a Hidden Digital Signature onto an Audio Signal Using Psychoacoustic Masking

Signal Processing for Speech Applications - Part 2-1. Signal Processing For Speech Applications - Part 2

ZLS38500 Firmware for Handsfree Car Kits

SMS045 - DSP Systems in Practice. Lab 1 - Filter Design and Evaluation in MATLAB Due date: Thursday Nov 13, 2003

Comparison of Multirate two-channel Quadrature Mirror Filter Bank with FIR Filters Based Multiband Dynamic Range Control for audio

Digital Signal Processing

Design Of Multirate Linear Phase Decimation Filters For Oversampling Adcs

Electrical & Computer Engineering Technology

EMBEDDED DOPPLER ULTRASOUND SIGNAL PROCESSING USING FIELD PROGRAMMABLE GATE ARRAYS

Performance Analysis of Feedforward Adaptive Noise Canceller Using Nfxlms Algorithm

Digital Filtering: Realization

Advanced AD/DA converters. ΔΣ DACs. Overview. Motivations. System overview. Why ΔΣ DACs

Design of FIR Filter on FPGAs using IP cores

ELEC3104: Digital Signal Processing Session 1, 2013 LABORATORY 3: IMPULSE RESPONSE, FREQUENCY RESPONSE AND POLES/ZEROS OF SYSTEMS

Decoding a Signal in Noise

Digitally controlled Active Noise Reduction with integrated Speech Communication

Design Digital Non-Recursive FIR Filter by Using Exponential Window

Problem Point Value Your score Topic 1 28 Discrete-Time Filter Analysis 2 24 Improving Signal Quality 3 24 Filter Bank Design 4 24 Potpourri Total 100

System analysis and signal processing

Audio Restoration Based on DSP Tools

Problem Point Value Your score Topic 1 28 Discrete-Time Filter Analysis 2 24 Upconversion 3 30 Filter Design 4 18 Potpourri Total 100

Isolated Word Recognition Based on Combination of Multiple Noise-Robust Techniques

ECE 5650/4650 Exam II November 20, 2018 Name:

A FEEDFORWARD ACTIVE NOISE CONTROL SYSTEM FOR DUCTS USING A PASSIVE SILENCER TO REDUCE ACOUSTIC FEEDBACK

Digital Signal Processing

Impulsive Noise Reduction Method Based on Clipping and Adaptive Filters in AWGN Channel

Speech Synthesis using Mel-Cepstral Coefficient Feature

Lecture 4 Biosignal Processing. Digital Signal Processing and Analysis in Biomedical Systems

THE problem of acoustic echo cancellation (AEC) was

Project 1. Notch filter Fig. 1: (Left) voice signal segment. (Right) segment corrupted by 700-Hz sinusoidal buzz.

Architecture design for Adaptive Noise Cancellation

Signal Processing Toolbox

ACOUSTIC feedback problems may occur in audio systems

Chapter 9. Chapter 9 275

Chapter 2: Digitization of Sound

Transcription:

Digital Signal Processing of Speech for the Hearing Impaired N. Magotra, F. Livingston, S. Savadatti, S. Kamath Texas Instruments Incorporated 12203 Southwest Freeway Stafford TX 77477 Abstract This paper presents some speech processing algorithms developed for hearing aid applications. However these algorithms are also applicable for other speech and audio applications. Considering that the basic properties of speech remain invariant across applications, it is logical to consider these algorithms under the broader umbrella of unified theory of speech. These algorithms have been implemented on Texas Instruments (TI) floating-point processor, the TMS320C3X (C3X). Portable C3X units were tested on hearing impaired subjects to assure the efficacy of the various speech processing algorithms. As these algorithms are being ported to TIs low-power fixed-point Digital Signal Processing (DSP) chips TMS320C54X and TMS320C55X, they are also being made XDAIS compliant. XDAIS is the acronym for express DSP Algorithm Standard. It is a software tool developed by TI which simplifies algorithm integration in today s complex DSP algorithm systems and also makes it easier to port algorithms across applications. Additionally we are performing quantization analysis of the algorithms to maintain the quality levels obtained via floating-point implementation. 1. Introduction The ultimate goal of this research is to facilitate the development of a fully programmable portable digital hearing aid that conditions the input speech signal based on environmental circumstances and the hearing impaired person's hearing loss - a fully customized digital hearing aid. Over the past several years, we have developed a prototype DSP Hearing Processor (DHP3X) system consisting of a laboratory-based PC system with a TMS320C3X DSP card and a wearable unit also based on the TMS320C3X DSP chip. The PC-based system is used to customize the hearing aid for a particular patient and the final algorithm is downloaded to a read-only memory (ROM) chip on the wearable unit [1]. These Algorithms are also being tested for other speech and audio applications. Since they are being developed using TI s XDAIS tool, they are easy to port across diverse applications. We are currently porting these algorithms to our fixed-point DSP chips the TMS320C54X and TMS320C55X chips (DHP5X). Typically, floating-point chips are larger and consume more power. However, it is usually easier to implement DSP algorithms on these chips since quantization effects are negligible for most applications. Fixed-point chips on the other hand tend to be smaller in size and have lower power requirements. However, the algorithms have to be analyzed for the effects of quantization noise on their performance and modified if necessary prior to porting to the fixed-point platform. The critical finite register length effects are [2]: errors due to A/D conversion, errors due to arithmetic roundoff, quantization of system parameters and bounds on the signal levels in order to prevent overflows. We are currently carrying out detailed quantization analysis of these algorithms. In the following discussion the DHP3X and DHP5X systems are collectively referred to as DHP.

2. Algorithms DHP3X is a binaural system capable of sampling two input microphone signals at sampling rates of up to 32 khz per channel. The minimum bandwidth of the hearing aid is 10 khz. At present we are researching three critical algorithms - Spectral Shaping (SS), Adaptive Noise Reduction (NR) and Multiband Dynamic range Compression (MDC). Figure 1 represents a palate of speech processing algorithms that could possibly be implemented on DHP because of its programmable design. A brief overview of the SS, NR and MDC algorithms is presented below. Frequency Shaping This signal processing stage shapes the speech spectrum (0 up to 16 khz) to compensate for the patient s hearing loss. The speech data is input to a binaural equalizer implemented as two banks of bandpass filters, one for each channel (ear). These filters are designed to have perfectly linear phase and high (>80 db) band isolation. The therapist can interactively (in real-time) choose the number of filters in each bank and select their critical frequency characteristics namely, their cutoff frequencies and isolation between different frequency bands. Figure 2 shows the resulting magnitude plot tuned for a particular patient s hearing loss. Because of the design flexibility built into the SS algorithm, the same DHP unit can be reprogrammed for different languages [1]. Adaptive Noise Reduction The noise reduction algorithm used in the DHP has been designed to work with just one input data channel. Hence the DHP processes the two (right and left ears) single input-single output channels independently. For each channel, the input signal is first high pass filtered to compensate for the low frequency spectral tilt [3] in speech signals. The high pass filter is a simple first order infinite impulse response (IIR) filter with tunable cutoff frequency. The noise reduction algorithm is referred to as Realtime Adaptive Correlation Enhancer (RACE). RACE is essentially an adaptive finite impulse response (FIR) filter. The autocorrelation coefficients of the input speech are used to update the RACE (FIR) coefficients after applying suitable gain control [4]. Figure 3 shows a generic block diagram for the RACE algorithm. The autocorrelation function is estimated using a recursive estimator represented by the equation below ^ ^ R xx ( n, = β R xx ( n 1, + (1 β ) x( n) x( n + where R^xx(n, and R^xx(n-1, are the autocorrelation estimates of the input x(n) at lag value k and at times n and n-1 respectively and β is a smoothing constant whose value lies between 0 and 1. The autocorrelation coefficients are estimated for lag values of k ranging from L to +L where L is known as the maximum lag. This results in a set of (2L+1) autocorrelation coefficients. Typically L is chosen to lie between 5 and 7. Both stability and rate of convergence are dependent on a single parameter, namely the smoothing constant β. A large value of β implies slow adaptation while a small value of β implies faster adaptation. Typically we set its value to correspond to a time constant of 4 msec for a 20 khz sampling rate. This is less than the short term stationarity of speech which is generally between 5-20 msec. Hence the values of β and L which determine the length of the adaptive FIR filter, should be chosen, so as not to exceed the short term stationarity assumption of speech [3].

Figure 4 shows the Power Density Spectra (PDS) corresponding to three time traces a clean sinusoid, a signal (input to RACE) consisting of the sinusoid corrupted with additive White Gaussion Noise (AWGN), and the RACE output signal for this particular input. From the figure it is evident that RACE reduces the amount of background white noise considerably. The noise floor is reduced by approximately 17 db. Similar tests conducted using CUNY nonsense syllables yielded SNR improvements on the order of 6 db. Amplitude Compression In addition to frequency shaping and noise suppression, the platform also permits real time implementation of multiband amplitude compression. Speech amplitude compression is essentially the task of controlling the overall gain of a speech amplification system. It essentially "maps" the dynamic range of the acoustic environment to the restricted dynamic range of the hearing impaired listener. Amplitude compression is achieved by applying a gain of less than one to a signal whenever its power exceeds a predetermined threshold [6]. Amplitude compression is based on the average power in the signal. The time constant of power estimation is used to modify the attack/release time of the compression algorithm [6]. If x(n) is the discrete time input signal, its estimated power p(n) is given by, p( n) = βp( n 1) + (1 β ) x 2 ( n) As long as the input power, to the compressor, is less than a preset threshold no compression takes place and the input is equal to the output. When the input power exceeds the threshold, the signal is attenuated. Once amplitude compression is being applied, if the attenuated input power exceeds a specified saturation power level the output power is held at a constant (saturated) level. The simulation data plots shown in Figure 5 demonstrate the efficiency of the compression alogrithm described above. The top trace in Figure 5 corresponds to the nonsense syllable 'za'. The lower plot in this figure represents the corresponding compressed output with a lower threshold of 0.15. Note that there is no 'spiking' in the time domain - a common artifact in amplitude compressed signals. Typically these spikes are representative of the compression algorithm missing the first few cycles of the loud sound. However, the power estimation employed in our implementation circumvents this problem. 3. Conclusions and Future Work The research summarized in this paper indicates that we can successfully implement various speech processing algorithms on Texas Instruments family of DSP chips floating-point as well as fixed-point. The results indicate that we can do so without sacrificing speech output quality. This is quite critical since the TMS320C55X chip is on target to be low enough in power consumption and small enough in footprint, to be used in developing true digital hearing aids that can be totally customized to individual hearing impaired subjects. Keeping in mind that these algorithms can be ported for other applications, they are being developed using TIs XDAIS tools. An added benefit of this portable prototyping device is that the efficacy of various speech processing strategies, including the algoritms described in this paper, can be measured using test subjects under real world conditions.

REFERENCES [1] N. Magotra, S. Sirivara, Real Time Digital Speech Processing Strategies for the hearing impaired, ICASSP 97, Munich, Germany. [2] C. T. Mullis, R. A. Roberts, Digital Signal Processing, Addison Wesley, 1987. [3] L. R. Rabiner, R. W. Schafer, Digital Processing of Speech Signals, Prentice-Hall, 1978. [4] N. Magotra, F. Livingston, S. Rajagopalan, "Single Channel Speech Enhancement in Real-time," 27 th Asilomar Conference on Signals, Systems and Computers, Asilomar, CA, Nov 1993. [5] S. Bangalore, "Quantization Analysis for Fixed-Point Implementation of Speech Processing for the Hearing Impaired," MS Thesis, EECE Department, University of New Mexico, 1999. [6]Sudheer Sirivara. Digital signal processing strategies for the hearing impaired. M.S Thesis, Electrical and Computer Engineering Department, University of New Mexico, July 1997.

STEREO INPUT (MICROPHONES) FREQUENCY SHAPING TMS320C3x/54x DSP ADAPTIVE NOISE REDUCTION INTERAURAL TIME DELAY TIMER MULTI CHANNEL AMPLITUDE COMPRESSION FUTURE ALGORITHM Figure 1 STEREO OUTPUT (EARPHONES) Figure 4 Figure 2 Figure 5 Autocorrelation Estimator u(n) y(n) Adaptive FIR Filter s(n) Figure 3