CO-CHANNEL SPEECH DETECTION APPROACHES USING CYCLOSTATIONARITY OR WAVELET TRANSFORM

Similar documents
Testing the Intelligibility of Corrupted Speech with an Automated Speech Recognition System

Enhancement of Speech Signal by Adaptation of Scales and Thresholds of Bionic Wavelet Transform Coefficients

Robust Voice Activity Detection Based on Discrete Wavelet. Transform

Detection, localization, and classification of power quality disturbances using discrete wavelet transform technique

Mel Spectrum Analysis of Speech Recognition using Single Microphone

Two-Feature Voiced/Unvoiced Classifier Using Wavelet Transform

Image Denoising Using Complex Framelets

Wavelet Speech Enhancement based on the Teager Energy Operator

STRUCTURE-BASED SPEECH CLASSIFCATION USING NON-LINEAR EMBEDDING TECHNIQUES. A Thesis Proposal Submitted to the Temple University Graduate Board

Different Approaches of Spectral Subtraction Method for Speech Enhancement

Introduction of Audio and Music

Fourier and Wavelets

Wavelet Transform. From C. Valens article, A Really Friendly Guide to Wavelets, 1999

Introduction to Wavelet Transform. Chapter 7 Instructor: Hossein Pourghassem

speech signal S(n). This involves a transformation of S(n) into another signal or a set of signals

Wavelet Transform. From C. Valens article, A Really Friendly Guide to Wavelets, 1999

Synchronous Overlap and Add of Spectra for Enhancement of Excitation in Artificial Bandwidth Extension of Speech

HTTP Compression for 1-D signal based on Multiresolution Analysis and Run length Encoding

Single Channel Speaker Segregation using Sinusoidal Residual Modeling

ON THE RELATIONSHIP BETWEEN INSTANTANEOUS FREQUENCY AND PITCH IN. 1 Introduction. Zied Mnasri 1, Hamid Amiri 1

Denoising Of Speech Signal By Classification Into Voiced, Unvoiced And Silence Region

Application of The Wavelet Transform In The Processing of Musical Signals

Application of Frequency-Shift Filtering to the Removal of Adjacent Channel Interference in VLF Communications

Epoch Extraction From Emotional Speech

HIGH QUALITY AUDIO CODING AT LOW BIT RATE USING WAVELET AND WAVELET PACKET TRANSFORM

Reduction of Musical Residual Noise Using Harmonic- Adapted-Median Filter

Speech Enhancement Using Spectral Flatness Measure Based Spectral Subtraction

Keywords Decomposition; Reconstruction; SNR; Speech signal; Super soft Thresholding.

Introduction to Wavelets Michael Phipps Vallary Bhopatkar

Classification of ships using autocorrelation technique for feature extraction of the underwater acoustic noise

Wavelet Transform for Bearing Faults Diagnosis

EE216B: VLSI Signal Processing. Wavelets. Prof. Dejan Marković Shortcomings of the Fourier Transform (FT)

Nonlinear Filtering in ECG Signal Denoising

Finite Word Length Effects on Two Integer Discrete Wavelet Transform Algorithms. Armein Z. R. Langi

TRANSFORMS / WAVELETS

A Method for Voiced/Unvoiced Classification of Noisy Speech by Analyzing Time-Domain Features of Spectrogram Image

A Survey and Evaluation of Voice Activity Detection Algorithms

Speech Enhancement using Wiener filtering

Epoch Extraction From Speech Signals K. Sri Rama Murty and B. Yegnanarayana, Senior Member, IEEE

Voiced/nonvoiced detection based on robustness of voiced epochs

ScienceDirect. Unsupervised Speech Segregation Using Pitch Information and Time Frequency Masking

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter

Voice Excited Lpc for Speech Compression by V/Uv Classification

Learning to Unlearn and Relearn Speech Signal Processing using Neural Networks: current and future perspectives

ARM BASED WAVELET TRANSFORM IMPLEMENTATION FOR EMBEDDED SYSTEM APPLİCATİONS

Audio Signal Compression using DCT and LPC Techniques

Spectral estimation using higher-lag autocorrelation coefficients with applications to speech recognition

Digital Audio Watermarking With Discrete Wavelet Transform Using Fibonacci Numbers

Orthonormal bases and tilings of the time-frequency plane for music processing Juan M. Vuletich *

Implementation of SYMLET Wavelets to Removal of Gaussian Additive Noise from Speech Signal

Target detection in side-scan sonar images: expert fusion reduces false alarms

8.3 Basic Parameters for Audio

Digital Image Processing

MODIFIED DCT BASED SPEECH ENHANCEMENT IN VEHICULAR ENVIRONMENTS

Performance study of Text-independent Speaker identification system using MFCC & IMFCC for Telephone and Microphone Speeches

SPEECH ENHANCEMENT USING PITCH DETECTION APPROACH FOR NOISY ENVIRONMENT

Encoding a Hidden Digital Signature onto an Audio Signal Using Psychoacoustic Masking

ROBUST PITCH TRACKING USING LINEAR REGRESSION OF THE PHASE

Voice Activity Detection for Speech Enhancement Applications

Speech Recognition using FIR Wiener Filter

Evoked Potentials (EPs)

RASTA-PLP SPEECH ANALYSIS. Aruna Bayya. Phil Kohn y TR December 1991

N J Exploitation of Cyclostationarity for Signal-Parameter Estimation and System Identification

Introduction to Wavelets. For sensor data processing

APPLICATION OF DISCRETE WAVELET TRANSFORM TO FAULT DETECTION

Determination of instants of significant excitation in speech using Hilbert envelope and group delay function

An Approach to Detect QRS Complex Using Backpropagation Neural Network

Aspiration Noise during Phonation: Synthesis, Analysis, and Pitch-Scale Modification. Daryush Mehta

Magnetic Tape Recorder Spectral Purity

Speech Synthesis; Pitch Detection and Vocoders

Electronic disguised voice identification based on Mel- Frequency Cepstral Coefficient analysis

Wavelet-based Voice Morphing

Applications of Music Processing

Estimation of Non-stationary Noise Power Spectrum using DWT

ScienceDirect. 1. Introduction. Available online at and nonlinear. c * IERI Procedia 4 (2013 )

Perception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 4: 7 Feb A. Faulkner.

Perception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 5: 12 Feb A. Faulkner.

INtroduction While the main focus of any speech recognition

Preeti Rao 2 nd CompMusicWorkshop, Istanbul 2012

Frequency Domain Analysis for Noise Suppression Using Spectral Processing Methods for Degraded Speech Signal in Speech Enhancement

NOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm

FAULT DETECTION OF FLIGHT CRITICAL SYSTEMS

THE APPLICATION WAVELET TRANSFORM ALGORITHM IN TESTING ADC EFFECTIVE NUMBER OF BITS

Calibration of Microphone Arrays for Improved Speech Recognition

ROBUST F0 ESTIMATION IN NOISY SPEECH SIGNALS USING SHIFT AUTOCORRELATION. Frank Kurth, Alessia Cornaggia-Urrigshardt and Sebastian Urrigshardt

Time-Frequency Enhancement Technique for Bevel Gear Fault Diagnosis

WAVELET OFDM WAVELET OFDM

Effects of Reverberation on Pitch, Onset/Offset, and Binaural Cues

Communications Theory and Engineering

Sound pressure level calculation methodology investigation of corona noise in AC substations

Singing Voice Detection. Applications of Music Processing. Singing Voice Detection. Singing Voice Detection. Singing Voice Detection

KONKANI SPEECH RECOGNITION USING HILBERT-HUANG TRANSFORM

Empirical Mode Decomposition: Theory & Applications

Audio Compression using the MLT and SPIHT

TIME FREQUENCY ANALYSIS OF TRANSIENT NVH PHENOMENA IN VEHICLES

Speech Compression Using Wavelet Transform

Instantaneous Baseline Damage Detection using a Low Power Guided Waves System

Wavelet Transform Based Islanding Characterization Method for Distributed Generation

Enhanced Waveform Interpolative Coding at 4 kbps

Transcription:

CO-CHANNEL SPEECH DETECTION APPROACHES USING CYCLOSTATIONARITY OR WAVELET TRANSFORM Arvind Raman Kizhanatham, Nishant Chandra, Robert E. Yantorno Temple University/ECE Dept. 2 th & Norris Streets, Philadelphia, PA 922-677, USA akizhana@astro.temple.edu, cnishant@astro.temple.edu, robert.yantorno@temple.edu http://www.temple.edu/speech_lab Stanley J. Wenndt Air Force Research Laboratory/IFEC, 32 Brooks Rd. Rome NY 344-454, USA wenndts@rl.af.mil ABSTRACT Co-channel speech occurs when one speaker s speech is corrupted by another speaker s speech. A co-channel detection system could provide information to suspend the operation of any speech processing system whose operation would be degraded if it were processing cochannel speech. In this paper we present two new methods of co-channel speech detection, one based on cyclostationarity and the other based on wavelet transform. Detection of co-channel speech in this paper refers to the detection of co-channel voiced speech, as it is not yet possible to detect unvoiced co-channel speech. Cyclostationary-based co-channel speech detection reveals that at least 65% of co-channel speech is correctly detected for different combinations of speech, e.g., male-male, female-female etc., with false alarms of approximately 24%. Investigation of the wavelet transform based co-channel speech detection reveals that at least 94% of co-channel speech is correctly detected with false alarms of approximately 28% making both methods tools for detecting cochannel speech.. INTRODUCTION Co-channel speech processing has been an area of interest and research for over three decades. Co-channel speech occurs when a speaker s speech (target) is degraded by another speaker s speech (interferer). Cochannel speech can occur in many situations, such as when two AM signals are transmitted on the same frequency, or when two people are speaking simultaneously (e.g. when talking on the telephone), or due to cross-talk from neighboring communication channels. An application of a co-channel detection system is shown in Figure. The co-channel detection system would provide information to suspend the operation of any speech processing system such as a speech recognition system, speaker identification system, etc., whose operation would be degraded if it were processing co-channel speech. Previously two approaches used for detection of co-channel speech have been presented; one was Spectral Autocorrelation Peak Valley Ratio (SAPVR) [], and the other used cepstral and pitch prediction features to determine the number of speakers in cochannel speech [2]. In this paper we present two new methods for detecting co-channel speech. Single or Co-channel Processing Technique Co-channel Detection System Suspend operation if co-channel speech detected Figure. Block Diagram of Application of Co-channel Detection System. A random signal can be modeled as cyclostationary, if its statistical parameters vary in time with single or multiple periodicities. The property of cyclostationarity of a signal can be exploited to detect the presence of signals buried in noise and/or severely masked by interference. This property makes this cyclostaionarity a good candidate for detection of co-channel speech. A wavelet transform uses localized basis functions, and hence is capable of yielding good signal approximations with very few terms of the wavelet transform. Because wavelets are localized within an interval, resolution in

time can be traded for resolution in frequency, making it feasible to investigate a particular signal interval efficiently. A wavelet transform can be used to detect the abrupt changes in the amplitude of the speech signal; this property makes this approach a good candidate for detection of co-channel speech. 2. CYCLOSTATIONARITY AND SPECTRAL REDUNDANCY A common assumption made by conventional statistical signal processing methods is that the random signals operated upon are stationary. That is, the parameters of the physical system that generates the random signal are time invariant. For most man-made signals, some parameters do vary periodically with time and in some cases harmonically unrelated periodicities are involved [3]. Investigations revealed that an inherent property of cyclostationarity signals is spectral redundancy, which corresponds to the correlation that exists between the random fluctuations of components of the signal residing in distinct spectral bands [3]. This property could be exploited to perform various signal processing tasks, such as: Detecting the presence of signals buried in noise and/or severely masked by interference. Recognizing such corrupted signals according to modulation type. Reduction of signal corruption due to cochannel interference and/or channel fading for single receiver systems. Linear periodic time variant prediction. is voiced [4]. A preset spectral flatness threshold (35dB) was found by performing a set of experiments. The spectral flatness of each frame of speech was compared against a preset threshold, and any frame above the threshold is determined to be voiced. 2. If the frame is determined to be voiced, a Hilbert transform is performed on the speech frame. 3. Then the convolution h*(n)*h(n+t) is performed on the Hilbert transformed signal. 4. The Fourier transform is performed on the output of convolution. From the output of the Fourier Transform three maxima are found. The time difference between the first and second maxima and the time difference between the second and third maxima is calculated. If both the differences are greater than a preset threshold (which was found to be 6 by performing a series of experiments and calculating the pitch difference), harmonic relations exist and therefore the frame is considered to be from a single speaker. 4 2-2 Co-channel Study using Cyclostationarity 5 5 2 25 3 35 4 45 5 x 8 co-channel speech, 52 samples = 32 msec 5 3. PROCEDURE FOR DETECTING CO-CHANNEL SPEECH USING CYCLOSTATIONARITY CCC 5 2 3 4 5 6 7 8 9 CCC points in msec The procedure for using the cyclostationary approach to detect co-channel speech is described in Figure 2. Voiced detection CCC Harmonic Relations Figure 2. Block Diagram of a Co-channel Detection System Using Cyclostationarity. CCC-Conjugated Cyclic Correlation Single Multi. Spectral flatness (SF) of the speech was used to determine if the speech frame under consideration Figure 3. Harmonic Structure for Single, Original Frame (top panel), Conjugated Cyclic Correlation (bottom panel). Figure 3 shows the original speaker speech frame in the top panel, conjugate cyclic correlation (CCC) of speech signal in the bottom panel. The time difference between the first and second maxima and the time difference between the second and third maxima in the bottom panel of Figure 3 is greater than the threshold (harmonic relations exist), hence it is a frame of single speaker.

Figure 4 shows the original speech frame in the top panel, conjugate cyclic correlation (CCC) of speech signal in the bottom panel. The time difference between the first and second maxima and the time difference between the second and third maxima in the bottom panel of Figure 3 is less than the threshold (no harmonic relations exist), hence it is a frame of cochannel speech. -2 2 CCC 6 5 4 3 2 Co-channel Study using Cyclostationarity 5 5 2 25 3 35 4 45 5 co-channel speech, 52 samples = 32 msec x 8 2 3 4 5 6 7 8 9 CCC points in msec Figure 4. Lack of Harmonic Structure of Cochannel, Original Frame (top panel), Conjugated Cyclic Correlation (bottom panel). 4. WAVELET ANALYSIS OF SPEECH SIGNALS The Fourier, sine and cosine transforms are non-local and hence there are limitations in time-frequency resolution of these transforms. Wavelets are small concentrated bursts of finite energy in time domain. A wavelet transform can be used to detect the abrupt changes in the amplitude of the speech signal, e.g., for pitch detection, [5] [6], or for voiced/unvoiced detection [6] [7], and is also a good candidate for detection of co-channel speech. If one observes a signal in a large window, gross features will be observed. However, small features are best observed by using small windows, which wavelet transforms can do. This allows the wavelets to reveal all the hidden features in the signal. This multi-resolution capability relies on being able to dilate (squeeze or/and expand) and translate the wavelet. Dyadic dilation (dilation by powers of 2) of the wavelet is the most popular of the wavelet functions and is also easy to implement [8] [9]. The wavelet prototype function used for analysis is called the mother wavelet [] [] [2] [3]. This function is dilated and translated to achieve the basis function at different scales. If x(t) is the signal and (t) is the wavelet function then a continuous wavelet transform (CWT) [CWT(b,a)] is a convolution of signal x(t) and wavelet function (t) expressed as: CWTx ( b, a) = x( t) Ψ *[( t b) / a] dt a () where a is the dilation parameter and b is the translation parameter. 5. PROCEDURE FOR DETECTING CO- CHANNEL SPEECH USING WAVELETS The procedure for using the wavelet approach to detect cochannel speech is described in Figure 5.. To determine the voiced speech, a discrete wavelet transform (DWT) is performed on the speech signal, on a frame-by-frame basis to obtain the approximate Voiced Detection WT Harmonic Relations Figure 5. Block Diagram of a Co-channel Detection System Using Wavelet Transform. WT-Wavelet Transform and detail coefficients. The complex continuous coefficients are obtained by computing CWT. Both DWT and CWT are used in order to ensure the reliability of voiced speech detection. 2. If the length of the frame is N, then for voiced speech the output of DWT and CWT will have about 9% of the total energy in the first N/2 samples and only about % of the total energy in the N/2+ to N samples. 3. For unvoiced speech, the output of DWT and CWT will not have 9% or more of the total energy for the first N/2 samples. 4. If a frame of speech is determined to be voiced, three maxima are found. The time difference between the first and second maxima and the time difference between the second and third maxima is calculated. If both the time differences are greater than a preset threshold (which was found to be 9 by performing series of experiments) harmonic relations exist and Single Multi

therefore the frame is considered to be from a single speaker. Figure 6 shows the original single speaker speech frame in the top panel, discrete wavelet transform (DWT) of speech signal in the middle panel and the harmonics of the discrete wavelet transformed speech signal in the bottom panel. Looking at the middle panel of Figure 5 we find that at least 9% of the energy of the output of the DWT is concentrated in first N/2 samples. Also, the time difference between the first and second maxima and the time difference between the second and third maxima in the bottom panel of Figure 6 is greater than the threshold (harmonic relations exist), hence it is a frame of single speaker speech. - - Co-channel Study Using Wavelets 5 5 2 25 3 35 4 45 5 Co-channel speech - 5 5 2 25 3 35 4 45 5 discrete wavelet transform msec 2 3 4 5 6 7 8 9 DWT points in msec Figure 6. Harmonic Structure for a Single, Original Frame (top panel), Discrete Wavelet Transform (DWT) (middle panel), Lower Portion of the DWT of the Middle Panel Showing Harmonic Relations (bottom panel). Looking at the middle panel of Figure 7 we find that at least 9% of energy of the output of the DWT is concentrated in first N/2 samples but the time difference between the first and second maxima and the time difference between the second and third maxima in the bottom panel of Figure 3 is less than the threshold (no harmonic relations exist), hence it is a frame of cochannel speech. 2-2 Co-channel Study Using Wavelets 5 5 2 25 3 35 4 45 5 Co-channel speech -2 5 5 2 25 3 35 4 45 5 Discrete wavelet transform msec 2-2 2 3 4 5 6 7 8 9 DWT points msec Figure 7. Harmonic Structure for a cochannel speech, original speech frame (top panel), discrete wavelet transform (DWT) (middle panel), lower portion of the DWT of the middle panel showing lack of Harmonic Relations (bottom panel). 6. EXPERIMENTS AND RESULTS Ten speech signals (5 male, 5 female) were taken from the TIMIT database. For each experiment, speech signals from two different talkers were combined to form a composite speech signal having overall TIR (Target-to- Interferer Ratio) of db (equal energy). The speech signals were sampled at 6kHz and then down sampled to 8 khz. The frame size was 32ms. Three different sets of experiments (male-male, female-female, male-female) were performed. Care was taken to ensure that the false alarm rate of the cyclostationary method was as close as possible to the false alarm rate of the wavelet transform approach so that comparison of both methods was possible. The cyclostationary-based co-channel detection system uses the procedure described in Section 3 to detect the existence of co-channel speech. As described in section 3, spectral flatness is performed on the input speech signal to detect voiced speech. For female-female speech after ten experiments 59.9% of co-channel speech was detected correctly and 42.% was missed, with false alarms of 23.%. For male-male speech 67.8% of co-channel speech was detected correctly and 32.2% missed, with false alarms of 22.3%. Similar experiments were performed for female-male speech and the results are presented in Table. Figure 8 shows the detection of co-channel speech based on cyclostationarity for female-female speech. The grey segments in Figure 8 are single speaker speech, the black

segments are the co-channel speech, and the rectangles are the detected co-channel speech segments using cyclostationarity..8.6.4.2 -.2 -.4 Cochannel Study - 27-Feb-22.8.6.4.2 -.2 -.4 -.6 -.8 - Cochannel Study - 27-Feb-22 TIRThreshold = 2 db - Frame length = 52 Singlespeakerspeech CoChannel Detected Cochannelspeech -.6 -.8 - TIRThreshold = 2 db - Frame length = 52 Single s Co-channel Detected Co-channel.5.5 2 2.5 3 3.5 Sample Number x 4 Figure 8. Detection of Co-channel Using Cyclostationarity, Single s (gray), Co-channel (black), Detected Co-channel (rectangles). The wavelet-based co-channel detection system uses the procedure described in Section 5 to detect the existence of co-channel speech. As described in section 5, DWT and CWT are performed on the input speech signal to detect voiced speech. Three maximas (peaks) are then found and the time differences between the maximas are compared to a preset threshold to detect the existence of co-channel speech. Figure 9 shows the detection of cochannel speech of female-female speech using wavelets. The grey segments in Figure 9 are single speaker speech, the black segments are the co-channel speech, and the rectangles are the detected co-channel speech segments using wavelet transform. For female-female speech we determined, after ten experiments, that 95. % of co-channel speech was detected correctly and 5.% were missed, with false alarms of 26.%. For male-male speech, 93.6% was determined as being correct and 6.4% was missed, with 27.2% false alarms. Similar sets of experiments were performed for female-male speech and results are tabulated in Table below..5.5 2 2.5 3 3.5 Sample Number x 4 Figure 9. Detection of Co-channel Using Wavelets, Single s (black), Co-channel (gray), Detected Co-channel (rectangles). Table: Results of Cyclostationary and Wavelets Based Co-channel Detection Systems. Co-channel speech % Correct % False Cyc Wav Cyc Wav Female-Female 59.9 95. 23. 26. Male-Male 67.8 93.6 22.3 27.2 Female-Male 66.7 94. 26. 29.6 Average 64.8 94.2 23.8 27.6 Cyc-cyclostationary, Wav- wavelets As can be observed from Table, for female-female speech, wavelets-based co-channel detection system gives a higher percentage correct as compared with the cyclostationary-based co-channel detection system. Similar observations can be made for male-male speech and female-male speech from Table. 7. SUMMARY In this paper we have presented two new methods of detecting co-channel speech, one based on cyclostationarity and the other based on wavelets. The results of our investigation of the cyclostationary and the wavelet methods of co-channel detection reveal that both systems can be used for detecting co-channel speech. The wavelet approach has an advantage over cyclostationary method in that it gives higher percent correct detection

when compared to cyclostationary, without a corresponding increase in false alarms. 8. FUTURE AREAS OF RESEARCH The possibilities of fusing the cyclostationary method and the wavelets method to determine a better cochannel speech detection system could also be explored. Further, the properties of statistical shape analysis and modulation maps could be studied in detail to explore the possibility of using them as methods of detecting co-channel speech. ACKNOWLEDGEMENT This effort was sponsored by the Air Force Research Laboratory, Air Force Material Command, and USAF, under agreement number F362-2-2-5. The U.S. Government is authorized to reproduce and distribute reprints for Government purposes notwithstanding any copyright annotation thereon. DISCLAIMER The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily the official policies or endorsements, either expressed or implied, of the Air Force Research Laboratory, or the U.S. Government. 9. REFERENCES [] Yantorno, R. E. A Study of Spectral Auto- Correlation Peak Value Ratio (SAPVR) as Method for Identification of Usable and Detection of Co-channel, Intelligent Signal Processing Workshop, Hungary, pp: 93-97, May 2. [5] Gonzalez, N. and Docampo, D. Application of Singularity Detection with Wavelets for Pitch Estimation of Signals, Proc. EUSIPCO, pp: 657-66, 994. [6] Janer, L. New Pitch Detection Algorithm Based on Wavelet Transform, IEEE-SP, pp: 65-68, 998. [7] Nam, H., Kim, H. and Yang, S. Verification Using Hybrid Model with Pitch Detection By Wavelets, Proc. IEEE ICASSP, pp: 53:56, 998. [8] Kadambe, S. and Boudreaux-Bartels, G.F. A Comparison of A Wavelet Functions for Pitch Detection of Signals, ICASSP, vol., pp: 449-452, May 99. [9] Johnson, I. A. Discrete Wavelet Transform Techniques in Processing, IEEE TENCON, pp: 54-59, 996. [] Davenport, M. R. and Garudadri, H. A Neural Net Acoustic Phonetic Feature Extraction Based on Wavelets, IEEE- Computers and Signal Processing, pp: 449-452, 99. [] Daubechies, I. Orthonormal Basis of Compactly Supported Wavelets, Comm. on Pure and Appl. Math. vol.4, pp: 99-996, Nov988. [2] Kaisheng, Y. and Zhigang, C. A Robust Feature- Perceptive Scalogram Based on Wavelet Analysis, ICSP, pp: 662-665, 998. [3] Pinter, I. Perceptual Wavelet-Representation of Signals, Computer, and Language, pp: -22, 996. [2] Lewis, A. and Ramachandran, P., On the Use of Cepstral and Pitch Prediction Features for Count Labelling Of Co-channel, ICSPAT Conference Proceedings, Toronto, Canada, pp: 2-24, Sep3-6, 998. [3] Gardner, Exploitation of spectral redundancy in cyclostationary signals, Signal Processing, pp: 4- -36, April 99. [4] Johnston, J. Transform Coding of Audio Signals Using Perceptual Noise Criteria IEEE J. on Select Areas in Comm., vol. SAC-6, pp: 34-323, 988.