Robust Low-Resource Sound Localization in Correlated Noise

Similar documents
Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter

Automotive three-microphone voice activity detector and noise-canceller

Students: Avihay Barazany Royi Levy Supervisor: Kuti Avargel In Association with: Zoran, Haifa

Localization of underwater moving sound source based on time delay estimation using hydrophone array

Different Approaches of Spectral Subtraction Method for Speech Enhancement

SPECTRAL COMBINING FOR MICROPHONE DIVERSITY SYSTEMS

Reduction of Musical Residual Noise Using Harmonic- Adapted-Median Filter

NOISE ESTIMATION IN A SINGLE CHANNEL

Sound Source Localization using HRTF database

Robust Speaker Identification for Meetings: UPC CLEAR 07 Meeting Room Evaluation System

IMPROVEMENT OF SPEECH SOURCE LOCALIZATION IN NOISY ENVIRONMENT USING OVERCOMPLETE RATIONAL-DILATION WAVELET TRANSFORMS

SOUND SOURCE LOCATION METHOD

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.

Airo Interantional Research Journal September, 2013 Volume II, ISSN:

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm

Reducing comb filtering on different musical instruments using time delay estimation

Chapter 4 SPEECH ENHANCEMENT

Enhancement of Speech Signal Based on Improved Minima Controlled Recursive Averaging and Independent Component Analysis

Direction-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method

ScienceDirect. Unsupervised Speech Segregation Using Pitch Information and Time Frequency Masking

ROBUST PITCH TRACKING USING LINEAR REGRESSION OF THE PHASE

Mel Spectrum Analysis of Speech Recognition using Single Microphone

High-speed Noise Cancellation with Microphone Array

Noise-robust compressed sensing method for superresolution

CHAPTER 3 SPEECH ENHANCEMENT ALGORITHMS

MODIFIED DCT BASED SPEECH ENHANCEMENT IN VEHICULAR ENVIRONMENTS

Recent Advances in Acoustic Signal Extraction and Dereverberation

RECENTLY, there has been an increasing interest in noisy

ROBUST echo cancellation requires a method for adjusting

Auditory System For a Mobile Robot

A MICROPHONE ARRAY INTERFACE FOR REAL-TIME INTERACTIVE MUSIC PERFORMANCE

Smart antenna for doa using music and esprit

RASTA-PLP SPEECH ANALYSIS. Aruna Bayya. Phil Kohn y TR December 1991

AdaBoost based EMD as a De-Noising Technique in Time Delay Estimation Application

Estimation of Non-stationary Noise Power Spectrum using DWT

WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY

Single channel noise reduction

Frequency Domain Analysis for Noise Suppression Using Spectral Processing Methods for Degraded Speech Signal in Speech Enhancement

Keywords Decomposition; Reconstruction; SNR; Speech signal; Super soft Thresholding.

MMSE STSA Based Techniques for Single channel Speech Enhancement Application Simit Shah 1, Roma Patel 2

The Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals

Audio Restoration Based on DSP Tools

SPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN. Yu Wang and Mike Brookes

Joint Position-Pitch Decomposition for Multi-Speaker Tracking

Isolated Word Recognition Based on Combination of Multiple Noise-Robust Techniques

Omnidirectional Sound Source Tracking Based on Sequential Updating Histogram

Speech Enhancement: Reduction of Additive Noise in the Digital Processing of Speech

Robust Voice Activity Detection Based on Discrete Wavelet. Transform

speech signal S(n). This involves a transformation of S(n) into another signal or a set of signals

Multiple Sound Sources Localization Using Energetic Analysis Method

Enhancement of Speech in Noisy Conditions

Applying the Filtered Back-Projection Method to Extract Signal at Specific Position

FPGA implementation of DWT for Audio Watermarking Application

IMPROVED COCKTAIL-PARTY PROCESSING

Signal segmentation and waveform characterization. Biosignal processing, S Autumn 2012

EXPERIMENTAL EVALUATION OF MODIFIED PHASE TRANSFORM FOR SOUND SOURCE DETECTION

BEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR

Acoustic Beamforming for Hearing Aids Using Multi Microphone Array by Designing Graphical User Interface

Quantification of glottal and voiced speech harmonicsto-noise ratios using cepstral-based estimation

Sound pressure level calculation methodology investigation of corona noise in AC substations

Broadband Signal Enhancement of Seismic Array Data: Application to Long-period Surface Waves and High-frequency Wavefields

Nonuniform multi level crossing for signal reconstruction

REAL-TIME BROADBAND NOISE REDUCTION

Speaker Localization in Noisy Environments Using Steered Response Voice Power

ONLINE REPET-SIM FOR REAL-TIME SPEECH ENHANCEMENT

Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor. Presented by Amir Kiperwas

Local Relative Transfer Function for Sound Source Localization

VHF Radar Target Detection in the Presence of Clutter *

Self Localization Using A Modulated Acoustic Chirp

Sound Processing Technologies for Realistic Sensations in Teleworking

Speech Enhancement using Wiener filtering

Speech Enhancement Techniques using Wiener Filter and Subspace Filter

Speech Signal Enhancement Techniques

SOUND FIELD MEASUREMENTS INSIDE A REVERBERANT ROOM BY MEANS OF A NEW 3D METHOD AND COMPARISON WITH FEM MODEL

Mikko Myllymäki and Tuomas Virtanen

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment

SYNTHETIC SPEECH DETECTION USING TEMPORAL MODULATION FEATURE

SPEECH ENHANCEMENT USING SPARSE CODE SHRINKAGE AND GLOBAL SOFT DECISION. Changkyu Choi, Seungho Choi, and Sang-Ryong Kim

A Weighted Least Squares Algorithm for Passive Localization in Multipath Scenarios

Non-intrusive intelligibility prediction for Mandarin speech in noise. Creative Commons: Attribution 3.0 Hong Kong License

Epoch Extraction From Emotional Speech

LONG RANGE SOUND SOURCE LOCALIZATION EXPERIMENTS

LOCALIZATION AND IDENTIFICATION OF PERSONS AND AMBIENT NOISE SOURCES VIA ACOUSTIC SCENE ANALYSIS

Spectral estimation using higher-lag autocorrelation coefficients with applications to speech recognition

AN ANALYSIS OF SPEECH RECOGNITION PERFORMANCE BASED UPON NETWORK LAYERS AND TRANSFER FUNCTIONS

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming

Improving Meetings with Microphone Array Algorithms. Ivan Tashev Microsoft Research

CHAPTER 4 VOICE ACTIVITY DETECTION ALGORITHMS

Single Channel Speaker Segregation using Sinusoidal Residual Modeling

Chapter 3. Source signals. 3.1 Full-range cross-correlation of time-domain signals

EE482: Digital Signal Processing Applications

27th Seismic Research Review: Ground-Based Nuclear Explosion Monitoring Technologies

Evaluating Real-time Audio Localization Algorithms for Artificial Audition in Robotics

A COHERENCE-BASED ALGORITHM FOR NOISE REDUCTION IN DUAL-MICROPHONE APPLICATIONS

Effects of Reverberation on Pitch, Onset/Offset, and Binaural Cues

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

Dominant Voiced Speech Segregation Using Onset Offset Detection and IBM Based Segmentation

Speech Enhancement Based On Noise Reduction

Wavelet Speech Enhancement based on the Teager Energy Operator

Auditory modelling for speech processing in the perceptual domain

Transcription:

INTERSPEECH 2014 Robust Low-Resource Sound Localization in Correlated Noise Lorin Netsch, Jacek Stachurski Texas Instruments, Inc. netsch@ti.com, jacek@ti.com Abstract In this paper we address the problem of sound source location using the time difference of arrival (TDOA) technique in an environment containing stationary correlated noise. We present a robust low-complexity method for enhancing estimation of sound direction, augmenting the well-known Generalized Cross-Correlation with Phase Transform (GCC- PHAT) approach. In the proposed method, the estimated cross-spectrum of a correlated background noise is subtracted from the observed spectrum. This effectively removes the phase distortion introduced by the interfering noise and significantly improves the robustness of the sound direction estimate. We test the performance of this approach on data collected and processed with a low-resource embedded platform. Results illustrate substantially enhanced performance over the baseline GCC-PHAT sound localization. Index Terms: sound source localization, source location, time-delay estimation, GCC-PHAT 1. Introduction Sound source localization is desirable in many environmentaware applications such as robotics, security, communications, as well as home or workplace management. The currently available sound-localization solutions typically require many sensors to be effective. To reduce cost and power consumption, the applications can benefit from limitedcapability low-resource sensor systems. The sound localization performed by each system could then be integrated within a larger sensor fusion process. This makes low-resource sound localization an attractive area of interest. Noise interference in each channel is often due to reverberation. However, in many applications the environment may contain interfering noise sources separate from the sound of interest, such as a fan in an office. Sensors are in practice also susceptible to interfering noise components such as electrical noise within the physical devices of the sensor system itself which may exhibit correlation. One commonly-used method for determining the location of a sound estimates the TDOA of a sound source relative to several microphones. A popular method for achieving this is the Generalized Cross-Correlation with Phase Transform (GCC-PHAT) method [1], which is attractive due to its low computational requirements and effectiveness in reverberant environments. A limitation of GCC-PHAT is that it obtains the TDOA estimate directly from the phase by normalizing the spectral magnitude at each frequency. This emphasizes the phase even at frequencies dominated by low-level background noise. Recently, many methods have been developed to deal with this issue. Some techniques modify the GCC-PHAT weighting function, for example by applying an SNRdependent exponent to the weighting function [2], adding a bias term in the denominator [3], using estimates of the phase statistics [3][4], or other SNR-based weighting functions [5]. In some methods, frequencies of low SNR have been temporally removed from consideration in the GCC-PHAT calculations [6]. Prior to applying GCC-PHAT, some methods reduce the effects of noise and remove unwanted signal components by performing spectral subtraction and mean normalization [3], or by decomposing the input using basis functions [7]. These techniques do not discriminate between correlated or uncorrelated noises, and thus may remove desired signal information. In [8], the phase of the noise signal in each channel is estimated during times when the desired signal is not present, and then the estimated signal without noise is generated prior to TDOA analysis. However, the noise signal phase is estimated for each channel individually without considering the noise signal correlations between channels. This paper presents a robust low-complexity method for enhancing estimation of sound direction by removing stationary interfering signals prior to the GCC-PHAT weighting and TDOA estimation. Our method is similar to [8], except that the estimation and noise removal is applied in the cross-spectral domain. This incurs little additional processing since the cross-spectrum is a required component of the GCC- PHAT processing. We test the method on data collected using a microphone pair connected to the processor in a Texas Instruments TMDSIPCAM8127J3 reference design IP camera, and illustrate significant interfering noise reduction and TDOA performance improvements. The paper is organized as follows. Section 2 presents the problem and formulates the proposed solution. Section 3 focuses on the practical implementation issues. Section 4 reviews the performed test and summarizes the results. We draw the conclusions in section 5. 2. Problem Statement This paper considers a two-microphone array in order to determine the sound direction based on the well-known time difference of arrival. We assume that the sound source is located far enough from the microphones that the far-field flatwavefront assumption applies. In addition, we assume that there exists some interfering noise signal that acts as a correlated signal in each channel. With these assumptions, we represent the observed signals at each microphone as: xi ( t) s( t si) b( t bi) i ( t) i 1,2 (1) Here x i (t) is the observed sound signal in microphone i at time t. We wish to determine the angular direction of the location of signal source s in relation to the two microphones using TDOA, where the different distances between the source s and the microphones i result in time delays si. The signal b is a background signal that appears as an interfering sound source, and which acts as a correlated noise with time delays of bi. The signals i represent uncorrelated noise in each channel due to effects such as electronic component thermal Copyright 2014 ISCA 2218 14-18 September 2014, Singapore

noise. We assume that s, b, and i come from different sources so they are uncorrelated with each other and are zero mean, and that b and i are stationary within the time of interest. 2.1. Proposed TDOA Direction Estimation Given the assumption of s, b, and i being uncorrelated with one another, the cross-correlation of the two microphone signals in (1) becomes Rx( ) x1 ( t) x2( t dt Rs ( s) Rb ( b) (2) where the time difference of arrival between microphones corresponding to the signal s is s s2 s1 and for the interfering signal b is b b2 b1. In the absence of the interfering noise b, the peak of the cross-correlation occurs at s. When noise is present, the crosscorrelation is a combination of the signal and the background noise, affecting the location of the estimated cross-correlation peak of R x( ). Since the background noise b is stationary, the problem may be practically mitigated by estimating the crosscorrelation component of the signal b during times when the signal s is not present, and subtracting it from the crosscorrelation of the observed signal. Rs ( s) Rx( Rb ( b) (3) One method to estimate the cross-correlation is to include voice activity detection (VAD) as in [8] to determine when only the interfering background signal is present. The estimate can be made from these periods. Once the estimated interfering background cross-correlation is removed, the resulting peak value should provide a better estimate of the TDOA of the signal s. 2.2. Spectral representation and PHAT The cross-correlation processing is often performed in the frequency domain by considering the signal cross-spectrum. This also allows the introduction of frequency weighting such as GCC-PHAT. Taking the Fourier transform of (2), we obtain: Gx( ( Gb ( 2 j 2 s j (4) S( e B( e b where G s ( is the cross-spectrum of the signal s, and G b( is the cross-spectrum of the interfering noise b. The time delays s and b are now reflected in phase shifts that are linear in frequency. In GCC-PHAT only the phase information contributes to the estimate of the time delay s. In (4) the term Gb( produces a phase error which causes the phase of the observed G x ( to differ from the phase of G s (. This phase error depends on the difference in phase between G s ( and G b(, as well as their relative magnitudes. For frequencies where G b( is greater than G s (, the phase of the interfering signal b dominates (even though the overall SNR may be high). To reduce the phase errors due to the interfering signal b, we want to perform the PHAT weighting, and to determine the TDOA estimate s, using the spectrum corresponding to the uncorrupted signal s, i.e. We can obtain the cross-spectrum G s( by estimating the interfering cross-spectrum G b( and subtracting it from G x (. For example, we can calculate an estimate of the cross-spectrum G b( during times when the signal s is absent. 3. Implementation In a practical implementation, we must perform processing on successive segments (frames) of the input signal. Processing in the frequency domain allows us to efficiently estimate the cross-spectrum of G b, and take advantage of frequency weighting of the cross-spectrum, such as GCC- PHAT. 3.1. Estimation of G b ( We begin by calculating the FFT of the observed signal for each frame. (For clarity in the following analysis we do not include a frame index in the equations.) Xi ( FFT( xi ( n)) i 1,2 (6) We next calculate the cross-spectrum of X i for the frame * G x ( X1 ( X2( (7) As mentioned above, we estimate the cross-spectrum G b ( during time periods when the speech signal s is not present. In the absence of s, the observed cross-spectrum will be a noisy representation of G b ( due to noises i. We average G b ( over a number of frames to reduce the effect of noise and produce an estimate such that: 2 j b Gˆ NT b( B( e (8) where B( is the spectrum of the interfering signal b, N is the size of the FFT, and T is the sample period of the signal. 3.2. Estimation of G s ( ( e j t s arg max d (5) t ( We obtain the signal cross-spectrum by removing the estimated interfering cross-spectrum: ( Gx( Gˆ b( (9) 2 j s S( e NT and apply the PHAT weighting to G s (. Ps ( ( ( The PHAT weighting removes all spectral amplitude information, so that the TDOA is reflected only in the phase information. Since the interfering phase information is reduced by (9), we expect improved estimation of s. 3.3. Estimation of s j s e NT (10) To estimate the delay s based on (5), a discrete implementation of the inverse Fourier transform is generated for each frame of data. For increased resolution, this is 2219

sometimes done by taking the inverse FFT of P s ( followed by interpolation to determine a more accurate estimate of the delay [6]. In our current method, we follow an approach outlined in [9]. Since the delay values of interest are limited to a range representing ±90 degrees relative to the microphone axis, we form a transformation matrix D for 2N +1 discrete test values spanning that range nd where n N N (11) cn where d is the distance between microphones and c is the speed of sound. The resolution of (and so the s resolution) can be adjusted by selection of N. Using the above values of, we form the matrix D of discrete exponential multipliers for frequency indices k: nd j j D( n, e NT e NT cn (12) with k 1 N 2 We apply this transformation to P s ( to produce the GCC- PHAT cross-correlation: R s ( n) real( D( n, Ps ( ) (13) k The delay estimate is determined by finding the index n of the largest value of R(n): nd ˆ nˆ arg max( Rs ( n)) and ˆ s (14) n cn The cross-spectrum of G s ( in (9) will be noisy due to the influence of the noises i. In the experiments described in Section 4, we test two methods to mitigate the effects of this noise. The first method averages R s (n) over a number of frames prior to calculating the estimate of s in (14). The second method averages the G s ( in (9) over a number of frames prior to applying the PHAT weighting in (10). 4. Experiments and Discussion We validate the proposed methods using a pair of microphones connected to a Texas Instruments TMDSIPCAM8127J3 reference design IP camera. The camera and microphones are mounted on a stand approximately 1m high and placed in an office room 3.7m by 2.5m. The microphone spacing is 10cm. A speaker, placed at the same height as the camera, is positioned in the room facing the center of the microphones, at a location that would result in a time delay of 0.28ms. The 16 khz speech signal played out of the speaker consisted of a concatenation of sentences spoken in a noise-free environment by several male and female subjects, having a total duration of about 69s. The nominal SNR recorded at the IP camera is 18dB. A plot of one microphone channel is shown in Figure 1. We use a frame size of 1024 samples, an FFT size N of 2048 (zero-padded), and we choose 90 as the value for N. For simplicity, instead of using VAD, we take the first 10 nonspeech frames to determine the correlated background noise (initial frames contain only background noise as can be seen from Figure 1). We use two different methods to mitigate the effects of the uncorrelated noise i. In one method we average the GCC- PHAT cross-correlation R s (n), and in the other we average the cross-spectrum G s (. In either case, the averages are performed over eight consecutive frames. Seconds Figure 1: Microphone channel data. For the first method, we determine G s ( with and without removing the estimated Gˆ noise as per (9). We apply the GCC-PHAT as in (10), calculate the cross-correlation R s (n) as in (13), and finally average the R s (n) values over consecutive frames. The results are shown in Figure 2. In Figure 2a we show a plot of R s (n) for each frame during the speech portion of the signal without removing Gˆ. The peak delays in R s (n) corresponding to 0.28ms can be seen. However there is interfering correlated noise around 0ms. Although the audio amplitude of the correlated noise at the beginning of the signal is quite small as seen in Figure 1, the contribution to the GCC- PHAT is often large enough to exceed the signal peak. Figure 2b is a scatter plot of the estimated delay GCC-PHAT values ˆ s as in (14) for each frame without removal of Gˆ. In Figure 2c we show the plot of R s (n) for each frame with prior removal of Gˆ. This results in a noticeable decrease in the interfering correlated noise, which allows for a much better detection of the signal delay. To confirm the effectiveness of the interference removal, in Figure 2d we show the scatter plot with removal of Gˆ. Comparing Figure 2b and Figure 2d, we can see that without the Gˆ removal many frames during speech have the peak locations at 0ms delay. With the Gˆ removal, there are fewer misidentified peak locations, demonstrating the effectiveness of the proposed solution. For the second method, we again determine G s ( with and without Gˆ as per (9). We now calculate the average the cross-spectra G s ( prior to applying the GCC-PHAT weighting of (10). Finally, we calculate the cross-correlation R s (n) as in (13), but do not average these (as done in the first method). The results are shown in Figure 3. In Figure 3a we show the plot of R s (n) during the speech portion of the signal without removing Gˆ. Again peaks can be seen at 0.28ms, along with the interfering peaks at 0ms. Comparing Figure 2a and Figure 3a, the peaks at 0.28ms appear more consistent and the noise peaks at 0ms are in many cases lower than the peaks at 0.28ms in Figure 3a. This is confirmed by the scatter plot of the estimated delay GCC-PHAT values ˆ s in Figure 3b, which show that there are fewer peaks around 0ms. With removal of Ĝ b, the plot of R s (n) is shown in Figure 3c. Again, the peaks at 0ms are reduced in amplitude. Comparing Figure 2c and Figure 3c, in the later the peaks at 0.28ms are more consistent and the interfering peaks at 0ms are less consistent in location and amplitude. Comparing the scatter plots of Figure 3b and Figure 3d, again we see correction of the peak locations. Comparing Figure 2d and Figure 3d, both methods produce similar improvements in location of the estimated signal peaks. This suggests choosing the averaging method 2220

(a) (a) (b) (b) (c) (c) (d) (d) Figure 2: Method 1: averaging over R s (n); a) R s (n) with interference; b) the corresponding values; c) R s (n) with interfering noise removed; d) the corresponding values. that reduces computation for best efficiency. The first method averages the 181 real values of R s (n) of the cross-correlation index, n over the frames, and the second method averages the 1024 complex values of G s ( of the cross-spectrum index, k. This favors the method of averaging of cross-correlation. In principle the solution presented can be used for multiple correlated interfering sound sources as long as they can be separated from the signal of interest, for example, by using VAD. The estimation of the interfering noise does not require a complex algorithm and will not require significant additional resources, since it can be developed in the process of the GCC-PHAT calculation. This makes it applicable to lowresource platforms. 5. Conclusions We presented a method to improve the TDOA estimate in the presence of a stationary interfering noise that exhibits correlation between microphone channels. We demonstrated that its contribution to phase distortion substantially affects the location of GCC-PHAT cross-correlation peaks, even though Figure 3: Method 2: averaging over ; a) R s (n) with interference; b) the corresponding values; c) R s (n) with interfering noise removed; d) the corresponding values. the interfering noise may be small in amplitude compared to the desired signal. We introduced a computationally efficient method to estimate and reduce this interfering phase distortion. The proposed solution effectively improves performance over baseline GCC-PHAT sound localization. In addition to reducing phase distortion due to the correlated noise, we compared two different methods of averaging to reduce the effects of the uncorrelated noise component. Experiments showed that without reducing the correlated noise distortion, averaging the cross-spectrum prior to the GCC-PHAT transformation provided more reliable estimates of TDOA. However, when we suppress the correlated noise interference, both cross-spectrum and crosscorrelation averaging methods yield comparable TDOA estimates. The reduced dimensionality of cross-correlation averaging is preferable for reduced computation. 2221

6. References [1] Knapp, C. H. and Carter, G. C. The Generalized Correlation Method for Estimation of Time, IEEE Trans. Acoustics, Speech, and Signal Processing, vol. 24, pp. 320 327, 1976. [2] Bo Qin, Heng Zhang, Qiang Fu, Yonghong Yan, Subsample Time Estimation via Improved GCC PHAT Algorithm, Proc. ICSP 2008, pp. 2979-2982, 2008. [3] Hong Liu and Miao Shen, Continuous Sound Source Localization based on Microphone Array for Mobile Robots, IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 4339-4339, 2010. [4] Bowon Lee, Amir Said, Ton Kalker, and Ronald W. Schafer, Maximum Likelihood Time Estimation with Phase Domain Analysis in the Generalized Cross Correlation work, Workshop on Hands-free Speech Communication and Microphone Arrays, pp. 89-92, 2008. [5] Valin, J. M., Michaud, F., Rouat, J., & Létourneau, D., Robust sound source localization using a microphone array on a mobile robot, IEEE/RSJ International Conference on Intelligent Robots and Systems, Vol. 2, pp. 1228-1233, 2003. [6] Stachurski, J., Netsch, L., & Cole, R., Sound source localization for video surveillance camera, IEEE 10th International Conference on Advanced Video and Signal Based Surveillance, pp. 93-98, 2013. [7] Wu, X., Jin, S., Zeng, Z., Xiao, Y., & Cao, Y., Location for audio signals based on empirical mode decomposition, IEEE International Conference on Automation and Logistics, pp. 1888-1891, 2009. [8] Athanasopoulos, G. & Verhelst, W., A phase-modified approach for TDE-based acoustic localization, Interspeech 2013, pp. 2890-2894, 2013. [9] Blandin, C., Ozerov, A., & Vincent, E., Multi-source TDOA estimation in reverberant audio using angular spectra and clustering, Signal Processing, 92(8), 1950-1960, 2012. 2222