Spectral contrast enhancement: Algorithms and comparisons q

Similar documents
A Digital Signal Processor for Musicians and Audiophiles Published on Monday, 09 February :54

HCS 7367 Speech Perception

Perception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 4: 7 Feb A. Faulkner.

Speech Enhancement Using Spectral Flatness Measure Based Spectral Subtraction

Perception of pitch. Importance of pitch: 2. mother hemp horse. scold. Definitions. Why is pitch important? AUDL4007: 11 Feb A. Faulkner.

REAL-TIME BROADBAND NOISE REDUCTION

Perception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 5: 12 Feb A. Faulkner.

Different Approaches of Spectral Subtraction Method for Speech Enhancement

CHAPTER 2 FIR ARCHITECTURE FOR THE FILTER BANK OF SPEECH PROCESSOR

You know about adding up waves, e.g. from two loudspeakers. AUDL 4007 Auditory Perception. Week 2½. Mathematical prelude: Adding up levels

Mel Spectrum Analysis of Speech Recognition using Single Microphone

Chapter 4 SPEECH ENHANCEMENT

Reading: Johnson Ch , Ch.5.5 (today); Liljencrants & Lindblom; Stevens (Tues) reminder: no class on Thursday.

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter

III. Publication III. c 2005 Toni Hirvonen.

speech signal S(n). This involves a transformation of S(n) into another signal or a set of signals

Quantification of glottal and voiced speech harmonicsto-noise ratios using cepstral-based estimation

AUDL GS08/GAV1 Auditory Perception. Envelope and temporal fine structure (TFS)

ME scope Application Note 01 The FFT, Leakage, and Windowing

Linguistic Phonetics. Spectral Analysis

Overview of Code Excited Linear Predictive Coder

Auditory Based Feature Vectors for Speech Recognition Systems

AN AUDITORILY MOTIVATED ANALYSIS METHOD FOR ROOM IMPULSE RESPONSES

WARPED FILTER DESIGN FOR THE BODY MODELING AND SOUND SYNTHESIS OF STRING INSTRUMENTS

Reducing comb filtering on different musical instruments using time delay estimation

RASTA-PLP SPEECH ANALYSIS. Aruna Bayya. Phil Kohn y TR December 1991

The role of intrinsic masker fluctuations on the spectral spread of masking

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.

Audio Signal Compression using DCT and LPC Techniques

Pre- and Post Ringing Of Impulse Response

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm

Can binary masks improve intelligibility?

The psychoacoustics of reverberation

EE482: Digital Signal Processing Applications

Perceptual Speech Enhancement Using Multi_band Spectral Attenuation Filter

IIR Ultra-Wideband Pulse Shaper Design

Auditory modelling for speech processing in the perceptual domain

FFT 1 /n octave analysis wavelet

DEEP LEARNING BASED AUTOMATIC VOLUME CONTROL AND LIMITER SYSTEM. Jun Yang (IEEE Senior Member), Philip Hilmes, Brian Adair, David W.

Enhancement of Speech Signal Based on Improved Minima Controlled Recursive Averaging and Independent Component Analysis

SGN Audio and Speech Processing

Psycho-acoustics (Sound characteristics, Masking, and Loudness)

Advanced audio analysis. Martin Gasser

DESIGN OF VOICE ALARM SYSTEMS FOR TRAFFIC TUNNELS: OPTIMISATION OF SPEECH INTELLIGIBILITY

ROBUST PITCH TRACKING USING LINEAR REGRESSION OF THE PHASE

An Adaptive Adjacent Channel Interference Cancellation Technique

A psychoacoustic-masking model to predict the perception of speech-like stimuli in noise q

Fundamental frequency estimation of speech signals using MUSIC algorithm

ScienceDirect. Unsupervised Speech Segregation Using Pitch Information and Time Frequency Masking

International Journal of Modern Trends in Engineering and Research e-issn No.: , Date: 2-4 July, 2015

Speech Signal Analysis

A Novel Technique or Blind Bandwidth Estimation of the Radio Communication Signal

EFFECTS OF PHYSICAL CONFIGURATIONS ON ANC HEADPHONE PERFORMANCE

Acoustics, signals & systems for audiology. Week 4. Signals through Systems

EC 554 Data Communications

Convention Paper Presented at the 112th Convention 2002 May Munich, Germany

Evaluation of clipping-noise suppression of stationary-noisy speech based on spectral compensation

ECE 556 BASICS OF DIGITAL SPEECH PROCESSING. Assıst.Prof.Dr. Selma ÖZAYDIN Spring Term-2017 Lecture 2

Comparison of Spectral Analysis Methods for Automatic Speech Recognition

Design of FIR Filters

Data and Computer Communications Chapter 3 Data Transmission

Signal Processing for Speech Applications - Part 2-1. Signal Processing For Speech Applications - Part 2

Proceedings of the 5th WSEAS Int. Conf. on SIGNAL, SPEECH and IMAGE PROCESSING, Corfu, Greece, August 17-19, 2005 (pp17-21)

Improving room acoustics at low frequencies with multiple loudspeakers and time based room correction

Structure of Speech. Physical acoustics Time-domain representation Frequency domain representation Sound shaping

EC209 - Improving Signal-To-Noise Ratio (SNR) for Optimizing Repeatable Auditory Brainstem Responses

IMPULSE RESPONSE MEASUREMENT WITH SINE SWEEPS AND AMPLITUDE MODULATION SCHEMES. Q. Meng, D. Sen, S. Wang and L. Hayes

FREQUENCY RESPONSE AND LATENCY OF MEMS MICROPHONES: THEORY AND PRACTICE

L19: Prosodic modification of speech

Classification of ships using autocorrelation technique for feature extraction of the underwater acoustic noise

Corso di DATI e SEGNALI BIOMEDICI 1. Carmelina Ruggiero Laboratorio MedInfo

Signals & Systems for Speech & Hearing. Week 6. Practical spectral analysis. Bandpass filters & filterbanks. Try this out on an old friend

FOURIER analysis is a well-known method for nonparametric

Digital Filters IIR (& Their Corresponding Analog Filters) Week Date Lecture Title

Multichannel level alignment, part I: Signals and methods

PIECEWISE LINEAR ITERATIVE COMPANDING TRANSFORM FOR PAPR REDUCTION IN MIMO OFDM SYSTEMS

Original Research Articles

Performance study of Text-independent Speaker identification system using MFCC & IMFCC for Telephone and Microphone Speeches

Modulation Domain Spectral Subtraction for Speech Enhancement

SPEECH AND SPECTRAL ANALYSIS

International Research Journal of Engineering and Technology (IRJET) e-issn: Volume: 03 Issue: 12 Dec p-issn:

SAMPLING THEORY. Representing continuous signals with discrete numbers

Speech Coding in the Frequency Domain

RECENTLY, there has been an increasing interest in noisy

Digitally controlled Active Noise Reduction with integrated Speech Communication

Radio Receiver Architectures and Analysis

Chapter 5 Window Functions. periodic with a period of N (number of samples). This is observed in table (3.1).

Design Of Multirate Linear Phase Decimation Filters For Oversampling Adcs

DERIVATION OF TRAPS IN AUDITORY DOMAIN

Pattern Recognition. Part 6: Bandwidth Extension. Gerhard Schmidt

FFT analysis in practice

Tone-in-noise detection: Observed discrepancies in spectral integration. Nicolas Le Goff a) Technische Universiteit Eindhoven, P.O.

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment

Block diagram of proposed general approach to automatic reduction of speech wave to lowinformation-rate signals.

REDUCING PAPR OF OFDM BASED WIRELESS SYSTEMS USING COMPANDING WITH CONVOLUTIONAL CODES

Dominant Voiced Speech Segregation Using Onset Offset Detection and IBM Based Segmentation

Sound pressure level calculation methodology investigation of corona noise in AC substations

Speech Synthesis using Mel-Cepstral Coefficient Feature

INTRODUCTION TO ACOUSTIC PHONETICS 2 Hilary Term, week 6 22 February 2006

Transcription:

Speech Communication 39 (2003) 33 46 www.elsevier.com/locate/specom Spectral contrast enhancement: Algorithms and comparisons q Jun Yang a, Fa-Long Luo b, *, Arye Nehorai c a Fortemedia Inc., 20111 Stevens Creek Boulevard, Suite 150, Cupertino, CA 95014, USA b Quicksilver Technology, 6640 Via Del Oro, San Jose, CA 95119, USA c ECE Department, University of Illinois at Chicago, 851 S. Morgan Street, 1120 SEO, Chicago, IL 60607, USA Abstract This paper investigates spectral contrast enhancement techniques and their implementation complexity. Three algorithms are dealt with in this paper. The first is the method described by Baer, Moore and Gatehouse. Two alternative methods are also proposed and investigated in this paper from a practical application and implementation point of view. Theoretical analyses and results from laboratory, simulation and subject listening show that spectral contrast enhancement and performance improvement can be achieved by use of these three methods with the appropriate selection of their relevant parameters. Ó 2002 Elsevier Science B.V. All rights reserved. Keywords: Signal processing; Noise reduction; Speech enhancement; Human audition; Auditory system; Real-time implementation 1. Introduction Spectral contrast is defined as the decibel difference between peaks and valleys in the spectrum. There are two general motivations behind spectral contrast enhancement for hearing-impaired (HI) people. First, in a sensorineural-impaired cochlea, auditory filters are generally broader than the normal and are in many cases abnormally asymmetrical. Processing through these abnormal filters may produce a smearing of spectral detail in the q The work of A. Nehorai was supported by the Air Force Office of Scientific Research under Grants F49620-99-1-0067 and F49620-00-1-0083, the National Science Foundation under Grant CCR-0105334, and the Office of Naval Research under Grant N00014-01-1-0681. The work of J. Yang and F.-L. Luo was conducted before they joined the companies listed above. * Corresponding author. E-mail address: falongl@yahoo.com (F.-L. Luo). internal representation of acoustic stimuli. Differences in amplitudes between peaks and valleys in the input spectrum may be reduced, making it more difficult to locate spectral prominence (i.e., formants) which provide crucial cues to speech intelligibility. To enhance spectral contrast may be of some help in compensating for the effects of this reduced frequency selectivity. Second, spectral analysis of speech in noise typically shows that these formants are well represented only when the input signal-to-noise ratio (SNR) is large enough but the spectral valleys between the formants are filled with noise. HI people have a reduced ability to pick out the spectral prominence, and are more affected by the noise filling in the valleys, partly because of their reduced frequency selectivity. Therefore, spectral contrast enhancement may be beneficial for the noise reduction. As a matter of fact, from a noise reduction point of view, spectral contrast enhancement can result in speech 0167-6393/02/$ - see front matter Ó 2002 Elsevier Science B.V. All rights reserved. PII: S0167-6393(02)00057-2

34 J. Yang et al. / Speech Communication 39 (2003) 33 46 enhancement in noise which is also useful for normal people. Spectral contrast enhancement has received intensive attention and a number of techniques have been proposed during the past two decades. The simplest idea, as proposed by Boers (1980), is to square spectrum levels and then to normalize amplitude. High-amplitude regions of the spectrum will grow more in amplitude when squared than will low-amplitude regions. Another method, proposed by Summerfield et al. (1985), is to decrease formant bandwidth for the synthesis. Narrowing these bandwidths leads to both sharper spectral peaks and greater peak-to-valley ratios (and hence, increased contrast). In the method of Bustamante and Braida (1986), contrast enhancement is based on principal components decomposition of short-term spectrum and made by inflating the amplitude of higher-order principal components that are most strongly associated with narrow-band features of spectral shape. Bunnell (1990) modified the spectrum using the following relation: T i ¼ CðS i SÞþS; ð1þ where T i is the target magnitude at frequency bin i, S i is the original magnitude at frequency bin i, S is the average spectrum level, and C is a contrast weight. All spectrum levels are in decibels. When C ¼ 1, the target envelope is the same as the original envelope. When C < 1, contrast reduction is produced. C > 1 produces contrast enhancement. Because the first four formants play the most important role, Finan and Liu (1994) proposed a linear prediction (LP) based formant enhancement technique. In this method, an allpole digital filter, determined by LP, was used to model the resonances of the vocal tract for each frame. The formants and their bandwidths were evaluated from the poles of the filter. FIR filters centered on the formants were used to enhance the first four formants. The outputs of the filters were summed together to give the enhanced frame of speech. The final enhanced speech signal was generated by rejoining the overlapping frames. Ribic et al. (1996) proposed another formantextraction based enhancement technique that extracts the first three formants and then modifies the spectrum values in frequency bins around these formants by designing appropriate contrast weights, which can be accomplished in either the frequency domain or the time domain. Simpson et al. (1990) described a method for increasing the difference in level between peaks and valleys in the spectrum which involves convolving the spectrum with a difference-of-gaussians (DoG) filter. This operation is similar to taking a smoothed second derivative of the spectrum. MooreÕs group at Cambridge University carried out a comprehensive investigation concerning the performance and implementation of this type of approach (Stone and Moore, 1992; Baer et al., 1993). However, despite these efforts and available methods, a good compromise involving quality and complexity in spectral contrast enhancement has not yet been reached, and more effort is highly desirable for the real-time implementation and real-world uses of spectral contrast enhancement techniques. For that reason, we further investigated the method described in (Stone and Moore, 1992; Baer et al., 1993) (for short, the method is hereafter referred as CambridgeÕs method) under various conditions such as different noise environments, different signal-to-noise ratios, changing the equivalent rectangular bandwidth (ERB) of auditory filters and the enhancement degree of DoG filters, with a specific frame configuration. Furthermore, from a practical application and implementation point of view, two alternative methods are also proposed. These two new proposed methods require much less computational complexity in comparison with CambridgeÕs method and make it possible to achieve real-time implementation. The rest of this paper is organized as follows. Section 2 presents a brief description of CambridgeÕs method and our simulation results under various conditions. Section 3 proposes a simple contrast enhancement algorithm and presents its combination with other processing in hearingsaid products along with illustrations of our results. Section 4 presents another proposed spectral contrast enhancement algorithm and the related results. In Section 5, we make further comparisons among three algorithms. Finally, we will offer some conclusions.

J. Yang et al. / Speech Communication 39 (2003) 33 46 35 2. Cambridge s method on speech contrast enhancement The schematic of the method proposed in (Baer et al., 1993) is shown in Fig. 1 which mainly consists of four steps: 1. Transform of the input signal to frequency domain by performing FFT. 2. Calculation of excitation pattern. This involves calculating the output of an array of simulated auditory filters in response to the magnitude spectrum. Each side of each auditory filter is modeled as an intensity-weighting function, assumed to have the form j wðf Þ¼ 1 þ p f f cj f c exp j p f f cj f c ; ð2þ where f c is the center frequency of the filter and p is a parameter determining the slope of the filter skirts. The value of p is assumed to be the same for the two sides of the filter. The ERB of these filters is 4f c =p. According to the definition of ERB (Moore and Gatehouse, 1983), we have p f f c f c 4ðf f c Þ ¼ f c ð0:00000623f c þ 0:09339Þþ28:52 : ð3þ The purpose of this step is to remove minor irregularities in the spectrum while preserving peaks corresponding to major spectral prominence in the speech. 3. Calculation of the enhanced magnitude spectrum. First, an enhancement function is derived from the above excitation pattern by a convolution-like process with a DoG function on an ERB scale. This DoG function is the sum of a positive Gaussian and a negative Gaussian that has twice the bandwidth of the positive Gaussian, that is,! DoGðf Þ¼p 1 ffiffiffiffiffi exp ðf f cþ 2 2p 2b 2!! 1 2 exp ðf f cþ 2 ; ð4þ 8b 2 where b is a parameter determining the bandwidth of the DoG function and is selected, as Baer et al. (1993) suggested, by using b ¼ k 2 ð0:00000623f 2 c þ 0:09339f c þ 28:52Þ, r ffiffiffiffiffiffiffiffiffiffi! 8ln2 3 ð5þ where k is an adjustable constant and whose selection will be discussed later. The details of this convolution-like process are described as follows: for a given center frequency of the DoG function, the value of the excitation pattern at each frequency (in linear power units) is multiplied by the value of the DoG function at the same frequency, and the products obtained in this way are summed. The magnitude value of the excitation pattern at that center frequency is then replaced by that sum. The enhancement function enðf Þ derived in the above convolution is then used to modify the excitation pattern. At each frequency where the enhancement function is positive, the excitation pattern is increased in magnitude; at frequency where the enhancement function is negative, the excitation pattern is decreased in magnitude. This can be achieved by the following operation: Fig. 1. Schematic diagram of CambridgeÕs method.

36 J. Yang et al. / Speech Communication 39 (2003) 33 46 spenðf Þ¼M enðf Þ logð enðf Þ jenðf Þj j jþþ logðexðf ÞÞ; ð6þ where spenðf Þ is the enhanced magnitude spectrum, exðf Þ is the input excitation pattern and M is a parameter which determines the degree of the enhancement. The first term on the right side of Eq. (6) is called a gain function. 4. The magnitude value spenðf Þ is expressed in linear amplitude units and then combined with the original phase values and finally the IFFT is used to obtain the processed speech. With a specific frame configuration, we investigated the performance of this scheme under various conditions: (1) Different signal and noise sources: speech in traffic noise, speech with water noise, speech in restaurant and cafeteria, speech in kitchen, speech with music, and so forth. (2) Different SNRs. SNR levels that we have dealt with are 15, 10, 5, 0, 5 and 10 db. (3) Different bandwidth parameter k: from 0.1 to 10. (4) Different enhancement degree parameter M: from 0.1 to 0.5. With these conditions, we made extensive simulation investigations. On the basis of informal subject listening, we then arrived at some conclusions. (1) This scheme becomes effective only in large SNR conditions (usually, larger than 10 db). Because this method does not identify the noise and the desired speech, processing the input with lower SNR in effect enhances noise rather than the desired speech. Figs. 2(a) (d) illustrate an example with (female) speech in traffic noise. The sampling rate is 16 000 Hz and the length of FFT and IFFT is 128. The duration of each input is about 20 s. In this example, we selected k ¼ 1 and M ¼ 0:1. Figs. 2(a) (d) correspond to SNRs of 0, 5, 10 and 15 db, respectively. It can be seen from these figures that the second processing unit in Fig. 1 first removes minor irregularities in the spectrum (similar to some kind of smooth processing) and then provides the excitation pattern. Main peaks in the excitation pattern are located at frequency 800, 1200 and 2050 Hz, respectively. The main valleys of the spectrum after the second step processing mainly are located at 1800, 2800 and 4500 Hz. In comparison with the excitation pattern curve of each figure, the spectrum of the system output is obviously sharpened to some extent, that is, the spectral contrast of the output has been enhanced. For example, the differences between the peak at 2050 Hz and the valley at 2800 Hz in the related spectra of Fig. 2(a) are 17 and 21 db before the enhancement processing and after the enhancement processing, respectively. In addition, the degree of the contrast enhancement depends on the SNR of the input signal, because excitation patterns and the enhancement gain functions will be different for different SNRs of input signal. As a matter of fact, the effect of SNR on the enhancement of the spectrum around 800 Hz is from the noise in this example. This finding also means that this scheme does not apply to the speech-like noise environments, especially in the case of low SNRs. It should be noted that the excitation pattern processing unit (Step 2) boosts the high-frequency part of the input signal. As a result, the system in Fig. 1 also could serve as a high-pass filter to some extent. (2) The width parameter k has a large effect on the enhancement result. Figs. 3(a) (e) illustrate a set of results for k at 0.1, 0.5, 1.0, 2.0 and 10.0, respectively. In this example, speech with water noise was considered and the SNR of the input is 0 db. The enhancement factor M is selected to be 0.1. Main peaks in the excitation pattern are located at frequency 1600, 2700 and 4300 Hz, respectively. The main valleys of the spectrum after the second step processing mainly are located at 2020 and 3200 Hz. The conflict between the signal distortion and spectral contrast enhancement exists in selecting the width parameter k. It is worth mentioning that the amount of spectral contrast enhancement decreases in the cases of both large and small values of k with the enhancement occurring mainly when k is around 1.0. This is mainly because the DoG function approaches a constant when k takes either large value or small value. To illustrate this effect, let us consider the spectral magnitude difference

J. Yang et al. / Speech Communication 39 (2003) 33 46 37 Fig. 2. The results of CambridgeÕs method with different SNRs. (a) SNR ¼ 0 db; (b) SNR ¼ 5 db; (c) SNR ¼ 10 db; (d) SNR ¼ 15 db. between the frequency bins 1600 and 2020 Hz. Their magnitude difference in the excitation pattern is 13 db. With the enhancement processing, the differences become 13.7, 15.2, 19.2, 19.8 and 13.8 db for k at 0.1, 0.5, 1.0, 2.0 and 10.0, respectively. Obviously, there is almost no enhancement for k ¼ 0:1 or 10.0 and there is a significant enhancement for k ¼ 1:0 and 2.0. However, it would be difficult to give a relationship in numerical quantities between the enhanced amount and the width parameter k. For simplicity, we generally select k ¼ 1:0, which means that the width of the positive lobe (between the zero-crossing points) of the DoG function equals the ERB of the auditory filter with the same center frequency. (3) The enhancement degree parameter M is another important factor that affects the output signal. It is easy to see from Eq. (6) that the enhancement amount of the spectral contrast is a monotonically increasing function of the parameter M. Although a large value of M will result in large enhancement in spectral contrast, this would give rise to a distortion of the signal. On the contrary, a small value of M will not distort the signal but will carry the cost of low enhancement and poor quality improvement. Figs. 4(a) (c) show a set of simulation results with M being 0.1, 0.3 and 0.5, respectively. In this simulation, male speech in restaurant noise (a speech-like noise) is considered and the SNR is 5 db. All our simulations show that the appropriate value of M is about 0.1, and it would be better to allow this value to be adjustable rather than fixed in hardware implementation.

38 J. Yang et al. / Speech Communication 39 (2003) 33 46 Fig. 3. The results of CambridgeÕs method with different bandwidth parameter k. (a) k ¼ 0:1; (b) k ¼ 0:5; (c) k ¼ 1:0; (d) k ¼ 2:0; (e) k ¼ 10:0. In summary, with appropriate selection of the related parameters, the effective enhancement of spectral contrast can be achieved by CambridgeÕs method. The key problem that this method suffers from is the extensive computational complexity, which we will deal with in Sections 3 and 5.

J. Yang et al. / Speech Communication 39 (2003) 33 46 39 Fig. 4. The results of CambridgeÕs method with different enhancement degree parameter M. (a) M ¼ 0:1; (b) M ¼ 0:3; (c) M ¼ 0:5. 3. A simple spectral contrast enhancement technique Although CambridgeÕs method can improve performance, the computational complexity involved in this method is very extensive. Steps 2 and 3 both involve convolution-like computation in the frequency domain, which makes it difficult to implement in real time. Based on this problem, we will propose a simple spectral contrast enhancement (SSCE) technique which includes the following steps: 1. Transform of the input signal to frequency domain by performing FFT. 2. Calculation of the enhancement magnitude spectrum by spoutðf Þ ¼ M logðspinðf ÞÞ þ logðspinðf ÞÞ; ð7þ where spinðf Þ is the magnitude spectrum obtained by Step 1 and M is the enhancement factor with positive value. 3. Generation of processed speech by taking the magnitude value spoutðf Þ, expressed in linear amplitude units, combined with the original phase values, and finally using IFFT. In comparison with CambridgeÕs method, this proposed method avoids the calculation of excitation pattern exðf Þ and the enhancement function enðf Þ which are the major burden of CambridgeÕs method and hence does not need any convolutionlike computation at all. As a result, this proposed method is very simple from a computation complexity and real-time implementation point of view. Now, we will prove that this proposed method can effectively enhance the spectral contrast.

40 J. Yang et al. / Speech Communication 39 (2003) 33 46 Assuming spinðf 1 Þ and spinðf 2 Þ are the magnitude spectrum at the peak and valley, respectively, then we have jspoutðf 1 Þ spoutðf 2 Þj ¼ð1þMÞj logðspinðf 1 ÞÞ logðspinðf 2 ÞÞj: ð8þ Because M is a positive constant, Eq. (8) shows the enhancement of the spectral contrast with an amount of 1 þ M. It is worth mentioning that if we replace the contrast weight C of Eq. (1) by 1 þ M, then Eq. (1) becomes T i ¼ð1 þ MÞS i MS; ð9þ which is similar to Eq. (7). The major difference between Eqs. (9) and (7) is the involvement of the average spectrum level S in Eq. (9), which requires additional computation. Moreover, as pointed out by Bunnell (1990), for the purpose of getting the desired performance improvement and overcoming disadvantages of his algorithm, non-uniform contrast weights should be used in his algorithm, that is, contrasts were enhanced mainly at middle frequencies, leaving high and low frequencies relatively unaffected. All these make the real-time implementation of BunnellÕs algorithm more complicated and more difficult than that of our proposed algorithm. In addition, Bunnell (1990) dealt with speech in quiet, rather than speech in noise; that is, the results of BunnellÕs algorithm for processing speech in noise were not reported. It should also be noted that if we choose M ¼ 1 in Eq. (7) then the processing of this proposed algorithm becomes to square spectrum as Boers used in (Boers, 1980). However, our experimental results have shown that M ¼ 1 always results in an unacceptable signal distortion, although this choice of enhancement degree parameter can offer the simplest implementation structure. As we can see from the following results, M should be less than 0.5 for real applications of this proposed algorithm. Because of its simplicity, we can implement this proposed algorithm in hardware and include it in the DSP based digital hearing-aid products. This hearing-aid system with this technique may include the following parts: A/D converter, window overlap, FFT, compression gain calculation, spectral contrast enhancement gain calculation, IFFT, overlap-add, and D/A converter. The A/D unit converts the microphone signal to the digital domain and then sends it to one programmable (assembly code) DSP chip which performs all the above processing. Hanning window overlap processing before FFT is necessary to overcome the time aliasing problem and its artifacts at the final output. FFT processing provides both the magnitude and phase in linear domain for each frequency bin. Compression gain part determines how large gain (amplification) would be used for each perceptual frequency band on the basis of the hearing loss characteristics (typically, the audiogram) by some available fitting algorithm. The contrast enhancement gain for each frequency bin is calculated according to the proposed algorithm. As a matter of fact, the contrast enhancement gain calculation is accomplished in linear domain mainly because all values in digital implementation are limited to the range from 1 to 1. After compression processing and contrast enhancement processing, the magnitude of each frequency bin is combined with the corresponding phase obtained by FFT and then IFFT processing is carried out. The overlap-add after IFFT is also necessary for undoing the processing of the window overlap before FFT. Finally the output is converted to acoustical signal via the receiver. In addition, because the output of the receiver may cause feedback to the microphone, an adaptive feedback cancellation processing should be included in this hearing-aid system. Investigation of the performance of this proposed algorithm has been made in various ways such as laboratory tests and subject listening (for both HI people and normal people), and under various conditions such as (1) different signal and noise sources: speech in traffic noise, speech with water noise, speech in restaurant and cafeteria, speech in kitchen, speech with music, and so forth. (2) different SNRs: 15, 10, 5, 0, 5 and 10 db. (3) different enhancement degree parameters M: from 0.1 to 0.5.

J. Yang et al. / Speech Communication 39 (2003) 33 46 41 Figs. 5(a) (d) are the results for speech in traffic noise and SNRs of 0, 5, 10 and 15 db, respectively. In this example, we selected M ¼ 0:3. Main peaks in the spectrum are located at frequency 810, 1180 and 2030 Hz, respectively; main valleys are located at 920, 1850 and 2880 Hz. Obviously, the differences between peaks and valleys have been enhanced by this proposed processing method. For example, the differences between the peak at 2030 Hz and the valley at 2880 Hz in the spectra of Fig. 5(a) are 25.8 and 33.9 db before and after enhancement processing, respectively. As a matter of fact, according to Eq. (8), the desired enhanced spectral difference is 33.6 db, which is approximated by the real result 33.9 db in this example. Table 1 gives a set of comparisons between the desired and measured enhancement amounts for different frequency bins in this example. Note that in this table, 500 Hz is taken as a reference frequency bin, the enhancement degree parameter M is selected to be 0.25, and the compression gain is 0 db for all frequency bins. Unlike CambridgeÕs method, the degree of the contrast enhancement in this proposed algorithm does not depend on the SNR of the input signal. This can been seen from Eq. (8) and the above results. However, the enhancement degree highly depends on values of the parameter M. To illustrate this, Figs. 6(a) (c) show a set of simulation Fig. 5. The results of the SSCE algorithm with different SNRs. (a) SNR ¼ 0 db; (b) SNR ¼ 5 db; (c) SNR ¼ 10 db; (d) SNR ¼ 15 db.

42 J. Yang et al. / Speech Communication 39 (2003) 33 46 Table 1 Enhancement comparisons between the desired and measured outputs db SPL 500 Hz 1000 Hz 1500 Hz 2000 Hz 2500 Hz 3000 Hz 3500 Hz 4000 Hz Input signal 50.7 70.8 79.1 59.2 63.2 65.2 58.7 54.3 Desired output 50.7 75.8 86.2 61.3 66.3 68.8 60.7 55.2 Measured output 50.7 75.4 86.3 61.1 66.8 69.1 60.3 55.5 Fig. 6. The results of the SSCE algorithm with different enhancement degree parameter M. (a) M ¼ 0:1; (b) M ¼ 0:3; (c) M ¼ 0:5. results for M at 0.1, 0.3 and 0.5, respectively. In this example, speech in restaurant noise is considered and the SNR is 5 db. These results and theoretical analyses demonstrate that this proposed method can enhance the spectral contrast effectively. However, this proposed method suffers from the same problem that CambridgeÕs method does, namely the conflict between the enhancement degree and distortion of the desired signal. In addition, this proposed method enhances all spectral details, even when they result from noise. Section 4 presents another method that avoids this problem by combining the above two methods: it is a modified version of CambridgeÕs method, using the enhancement method proposed in Section 3.

J. Yang et al. / Speech Communication 39 (2003) 33 46 43 4. Excitation pattern based method spoutðf Þ ¼ M logðexðf ÞÞ þ logðexðf ÞÞ; ð10þ As pointed out in Section 2, the calculation of excitation pattern exðf Þ is mainly used to avoid enhancing spectral details and to preserve peaks corresponding to major spectral prominence in the speech. With this, if we simply replace the magnitude spectrum spinðf Þ of Eq. (7) by the excitation pattern, then we can get another spectral contrast enhancement method as follows: 1. Transform of the input signal to frequency domain by performing FFT. 2. Calculation of excitation pattern exðf Þ by use of Eqs. (2) and (3). 3. Enhancement of the excitation pattern by use of which is based on Eq. (7). 4. Combination of the magnitude value spoutðf Þ expressed in linear amplitude units with the original phase values and obtaining the processed speech using IFFT. In comparison with CambridgeÕs method, this new method needs only the calculation of excitation pattern; hence its computational complexity is about half that of CambridgeÕs method. On the other hand, the cost to avoid enhancing spectral details is the calculation of excitation pattern, which makes this excitation-pattern based (EPB) method much more computationally expensive in Fig. 7. The results of the EPB algorithm with different SNRs. (a) SNR ¼ 0 db; (b) SNR ¼ 5 db; (c) SNR ¼ 10 db; (d) SNR ¼ 15 db.

44 J. Yang et al. / Speech Communication 39 (2003) 33 46 comparison with the proposed method in Section 3. Under the same situations and the same frame configuration as in Section 3, we investigated the performance of this EPB method. Figs. 7(a) (d) are the results with a speech in traffic noise and with SNRs of 0, 5, 10 and 15 db, respectively. In this example, we selected M ¼ 0:3. Note that the input signal in this example is the same as that of Fig. 2(a) (d). Because the enhancement gain function is no longer involved in this proposed method, the effect of the SNR of the input signal on the enhancement amount of the output spectrum is much less than that with CambridgeÕs method. However, the parameter M is still the principal factor determining the enhancement amount of the output spectrum. To illustrate it, Figs. 8(a) (c) show a set of simulation results with M being 0.1, 0.3 and 0.5, respectively. In this example, speech in restaurant noise is considered and the SNR is 5 db. 5. Comparisons and discussions ofthe three algorithms In this section, we will make further comparisons and discussions about these three spectral contrast enhancement algorithms with emphasis on their computational complexity and implementation. As mentioned in the above sections, CambridgeÕs algorithm is the most complicated from an implementation point of view. As a further comparison, Tables 2 4 give the number of multiplications, additions and coefficients (corresponding to data memory) required in each step Fig. 8. The results of the EPB algorithm with different enhancement parameter M. (a) M ¼ 0:1; (b) M ¼ 0:3; (c) M ¼ 0:5.

J. Yang et al. / Speech Communication 39 (2003) 33 46 45 Table 2 Complexity for implementing CambridgeÕs algorithm Multiplications Additions Additional data memory (Bytes) N Stage 1 2 2ðNÞ 11 N þ 1 7 8 4 2ðNÞ 3N þ 2 0 Stage 2 N 2 N 2 N 2N 2 Stage 3 N 2 þ 2N N 2 þ N 2 1 2N 2 Stage 4 N N 0 N Stage 5 2 2ðNÞ 9 N þ 2 7 8 4 2ðNÞ 5 N þ 1 4 0 Total 2N 2 þ N log 2 ðnþþ 1 N þ 3 2 2N 2 þ 7 N log 2 2ðNÞ 15 4 4N 2 Table 3 Complexity for implementing SSCE algorithm Multiplications Additions Additional data memory (Bytes) N Stage 1 2 2ðNÞ 11 7 8 4 2ðNÞ 3N þ 2 0 Stage 2 N N 0 N Stage 3 2 2ðNÞ 9 N þ 2 7 8 4 2ðNÞ 5 N þ 1 4 0 Total N log 2 ðnþ 3 N þ 3 7 2 2 2ðNÞ 13 N þ 3 4 0 Table 4 Complexity for implementing EPB algorithm Multiplications Additions Additional data memory (Bytes) N Stage 1 2 2ðNÞ 11 N þ 1 7 8 4 2ðNÞ 3N þ 2 0 Stage 2 N 2 N 2 N 2N 2 Stage 3 N N 0 N Stage 4 2 2ðNÞ 9 N þ 2 7 8 4 2ðNÞ 5 N þ 1 4 0 Total N 2 þ N log 2 ðnþ 3 N þ 3 2 N 2 þ 7 N log 2 2ðNÞ 17 N þ 3 4 2N 2 for implementing CambridgeÕs algorithm, the SSCE algorithm proposed in Section 3, and the EPB algorithm proposed in Section 4, respectively, where 2N is the length of FFT and IFFT. It should be noted that Stages 3 and 4 of Table 2 form Step 3 of CambridgeÕs algorithm. Also, in these tables N must be a power of 2 and greater than 4. Moreover, because the DFT of real-data has the conjugate symmetry, we can use length-n complex-valued FFT to calculate length-2n realvalued FFT and IFFT with some additional preprocessing and post-processing to further reduce the computational complexity (Guo et al., 1998). This reduction of complexity has been taken into account in the calculation related to numbers in these tables. In addition, there are two ways to get the weight coefficients required in CambridgeÕs algorithm and the EPB algorithm in real-time implementation. One is to calculate them according to Eq. (2) online, which does not need additional data memory to store these coefficients. The other is to first calculate them off-line and then to store them in additional data memory. Because there is a symmetric property in these coefficients, only half of coefficients need to be stored. In the results of Tables 2 and 4 this symmetry has been used. The DoG function required in CambridgeÕs algorithm has the same problem and the same property as above for weight coefficients, which has been taken into account in Table 2 as well. In these tables, we assume that one coefficient needs two bytes (16 bits). For the configuration with the frame length is 128, that is, N ¼ 64, Table 5 gives the main operation (multiplication and addition) number and data memory size for implementing the three algorithms. It can be seen from Table 5 that the number of main operations required in the SSCE

46 J. Yang et al. / Speech Communication 39 (2003) 33 46 Table 5 Complexity for three algorithms with N ¼ 64 Multiplications Additions Additional data memory (Bytes) CambridgeÕs algorithm 8611 9298 16 384 SSCE algorithm 291 1139 0 EPB algorithm 4387 5171 8192 algorithm is only one twelfth of that required in CambridgeÕs algorithm and one sixth of that required in the EPB algorithm. In addition, no additional data memory is needed at all in the SSCE algorithm. These properties of the SSCE algorithm provide great simplicity for implementing this algorithm in real time. 6. Conclusions This paper investigated spectral contrast enhancement techniques and their hardware implementation complexity. Because of the extensive computational complexity of CambridgeÕs method, we proposed two alternative methods and investigated their performance. We have implemented one of these two proposed algorithms in hardware. The theoretical analysis, laboratory testing and subject listening results have shown that by using these three methods, the desired spectral contrast enhancement and performance improvement can be achieved with the appropriate selection of the related parameters. The common problem of these three methods is the conflict between the enhancement of spectral contrast and the distortion of the desired signal. Consequently, we make a trade off and select the appropriate values of the related parameters according to application situations. Acknowledgements We are grateful to the anonymous reviewers for their very useful suggestions and valuable comments. References Baer, T., Moore, B.C.J., Gatehouse, S., 1993. Spectral contrast enhancement of speech in noise for listeners with sensorineural hearing impairment: effects on intelligibility, quality, and response times. J. Rehab. Res. Devlop. 30 (1), 49 72. Boers, P.M., 1980. Formant enhancement of speech for listeners with sensorineural hearing loss. In: IPO Annual Progress Report No.15, Institut voor Perceptie Onderzoek, The Netherlands, pp. 21 28. Bunnell, T.H., 1990. On enhancement of spectral contrast in speech for hearing-impaired listeners. J. Acoust. Soc. Amer. 88 (6), 2546 2556. Bustamante, D.K., Braida, L.D., 1986. Wideband compression and spectral sharpening for hearing-impaired listeners. J. Acoust. Soc. Amer. 80 (Suppl. 1), S12 S13. Finan, R.A., Liu, Y., 1994. Formant enhancement of speech for listeners with impaired frequency selectivity. Biomed. Eng., Appl. Basis Comm. 6 (1), 59 68. Guo, H., Sitton, G.A., Burrus, C.S., 1998. The quick Fourier transform: an FFT based on symmetries. IEEE Trans. Signal Process. 46 (2), 335 341. Moore, B.C.J., Gatehouse, S., 1983. Suggested formulae for calculating auditory filter bandwidths and excitation patterns. J. Acoust. Soc. Amer. 74 (3), 750 753. Ribic, Z., Yang, J., Latzel, M., 1996. Adaptive spectral contrast enhancement based on masking effect for the hearing impaired. In: Proc. 1996 IEEE Internat. Conf. on Acous. Speech and Signal Process. Conf., vol. 2, pp. 937 940. Simpson, A.M., Moore, B.C.J., Glasberg, B.R., 1990. Spectral enhancement to improve the intelligibility of speech in noise for hearing impaired listeners. Acta Otolaryngol. 469 (Suppl.), 101 107. Stone, M.A., Moore, B.C.J., 1992. Spectral feature enhancement for people with sensorineural hearing impairment: effects on speech intelligibility and quality. J. Rehab. Res. Develop. 29 (2), 39 56. Summerfield, Q., Foster, J., Tyler, R., 1985. Influences of formant bandwidth and auditory frequency selectivity on identification of place of articulation in stop consonants. Speech Communication 4, 213 229.