Sound Source Localization using HRTF database

Similar documents
HRIR Customization in the Median Plane via Principal Components Analysis

Sound Source Localization in Median Plane using Artificial Ear

Ivan Tashev Microsoft Research

Auditory Localization

ENHANCED PRECISION IN SOURCE LOCALIZATION BY USING 3D-INTENSITY ARRAY MODULE

A Virtual Audio Environment for Testing Dummy- Head HRTFs modeling Real Life Situations

Airo Interantional Research Journal September, 2013 Volume II, ISSN:

Acoustics Research Institute

Convention Paper Presented at the 139th Convention 2015 October 29 November 1 New York, USA

TDE-ILD-HRTF-Based 2D Whole-Plane Sound Source Localization Using Only Two Microphones and Source Counting

Enhancing 3D Audio Using Blind Bandwidth Extension

Proceedings of Meetings on Acoustics

Sound Processing Technologies for Realistic Sensations in Teleworking

Robust Low-Resource Sound Localization in Correlated Noise

Spatial Audio & The Vestibular System!

EE1.el3 (EEE1023): Electronics III. Acoustics lecture 20 Sound localisation. Dr Philip Jackson.

IMPROVED COCKTAIL-PARTY PROCESSING

Listening with Headphones

Convention Paper 9870 Presented at the 143 rd Convention 2017 October 18 21, New York, NY, USA

Computational Perception /785

Study on method of estimating direct arrival using monaural modulation sp. Author(s)Ando, Masaru; Morikawa, Daisuke; Uno

ORIENTATION IN SIMPLE VIRTUAL AUDITORY SPACE CREATED WITH MEASURED HRTF

Sound source localization and its use in multimedia applications

Multiple Sound Sources Localization Using Energetic Analysis Method

EFFECTS OF PHYSICAL CONFIGURATIONS ON ANC HEADPHONE PERFORMANCE

Binaural Hearing. Reading: Yost Ch. 12

WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY

Computational Perception. Sound localization 2

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming

27th Seismic Research Review: Ground-Based Nuclear Explosion Monitoring Technologies

ONE of the most common and robust beamforming algorithms

Holographic Measurement of the 3D Sound Field using Near-Field Scanning by Dave Logan, Wolfgang Klippel, Christian Bellmann, Daniel Knobloch

Microphone Array Design and Beamforming

ROOM AND CONCERT HALL ACOUSTICS MEASUREMENTS USING ARRAYS OF CAMERAS AND MICROPHONES

Introduction. 1.1 Surround sound

Psychoacoustic Cues in Room Size Perception

A binaural auditory model and applications to spatial sound evaluation

BEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR

Multi-channel Active Control of Axial Cooling Fan Noise

Sound Radiation Characteristic of a Shakuhachi with different Playing Techniques

PERSONALIZED HEAD RELATED TRANSFER FUNCTION MEASUREMENT AND VERIFICATION THROUGH SOUND LOCALIZATION RESOLUTION

Convention Paper Presented at the 125th Convention 2008 October 2 5 San Francisco, CA, USA

Binaural Speaker Recognition for Humanoid Robots

From Binaural Technology to Virtual Reality

Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks

Automotive three-microphone voice activity detector and noise-canceller

HRTF measurement on KEMAR manikin

Numerical Study of Stirring Effects in a Mode-Stirred Reverberation Chamber by using the Finite Difference Time Domain Simulation

Active Noise Cancellation System Using DSP Prosessor

FREQUENCY RESPONSE AND LATENCY OF MEMS MICROPHONES: THEORY AND PRACTICE

Automatic Text-Independent. Speaker. Recognition Approaches Using Binaural Inputs

SOPA version 2. Revised July SOPA project. September 21, Introduction 2. 2 Basic concept 3. 3 Capturing spatial audio 4

The psychoacoustics of reverberation

Omnidirectional Sound Source Tracking Based on Sequential Updating Histogram

Audio Engineering Society. Convention Paper. Presented at the 129th Convention 2010 November 4 7 San Francisco, CA, USA. Why Ambisonics Does Work

Michael Brandstein Darren Ward (Eds.) Microphone Arrays. Signal Processing Techniques and Applications. With 149 Figures. Springer

Reduction of Musical Residual Noise Using Harmonic- Adapted-Median Filter

Blind source separation and directional audio synthesis for binaural auralization of multiple sound sources using microphone array recordings

Simultaneous Recognition of Speech Commands by a Robot using a Small Microphone Array

Extracting the frequencies of the pinna spectral notches in measured head related impulse responses

3D Sound System with Horizontally Arranged Loudspeakers

Speaker Localization in Noisy Environments Using Steered Response Voice Power

Smart antenna for doa using music and esprit

Audio Restoration Based on DSP Tools

Localization of underwater moving sound source based on time delay estimation using hydrophone array

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.

UWB SHORT RANGE IMAGING

396 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 2, FEBRUARY 2011

SIMPLE GEAR SET DYNAMIC TRANSMISSION ERROR MEASUREMENTS

The Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals

A triangulation method for determining the perceptual center of the head for auditory stimuli

On distance dependence of pinna spectral patterns in head-related transfer functions

Proceedings of Meetings on Acoustics

Upper hemisphere sound localization using head-related transfer functions in the median plane and interaural differences

INVESTIGATING BINAURAL LOCALISATION ABILITIES FOR PROPOSING A STANDARDISED TESTING ENVIRONMENT FOR BINAURAL SYSTEMS

Envelopment and Small Room Acoustics

DESIGN OF VOICE ALARM SYSTEMS FOR TRAFFIC TUNNELS: OPTIMISATION OF SPEECH INTELLIGIBILITY

BINAURAL RECORDING SYSTEM AND SOUND MAP OF MALAGA

Measuring impulse responses containing complete spatial information ABSTRACT

Binaural Sound Localization Systems Based on Neural Approaches. Nick Rossenbach June 17, 2016

University of Huddersfield Repository

URBANA-CHAMPAIGN. CS 498PS Audio Computing Lab. 3D and Virtual Sound. Paris Smaragdis. paris.cs.illinois.

Analysis of Frontal Localization in Double Layered Loudspeaker Array System

Room Impulse Response Modeling in the Sub-2kHz Band using 3-D Rectangular Digital Waveguide Mesh

Different Approaches of Spectral Subtraction Method for Speech Enhancement

Binaural Hearing- Human Ability of Sound Source Localization

Binaural Sound Source Localization Based on Steered Beamformer with Spherical Scatterer

Subband Analysis of Time Delay Estimation in STFT Domain

Robotic Spatial Sound Localization and Its 3-D Sound Human Interface

Final Project: Sound Source Localization

3D sound image control by individualized parametric head-related transfer functions

The effects of the excitation source directivity on some room acoustic descriptors obtained from impulse response measurements

Source Localisation Mapping using Weighted Interaural Cross-Correlation

Calibration of Microphone Arrays for Improved Speech Recognition

Sound localization with multi-loudspeakers by usage of a coincident microphone array

ANALYZING NOTCH PATTERNS OF HEAD RELATED TRANSFER FUNCTIONS IN CIPIC AND SYMARE DATABASES. M. Shahnawaz, L. Bianchi, A. Sarti, S.

Interference in stimuli employed to assess masking by substitution. Bernt Christian Skottun. Ullevaalsalleen 4C Oslo. Norway

Proceedings of Meetings on Acoustics

Recent Advances in Acoustic Signal Extraction and Dereverberation

Underwater Wideband Source Localization Using the Interference Pattern Matching

Transcription:

ICCAS June -, KINTEX, Gyeonggi-Do, Korea Sound Source Localization using HRTF database Sungmok Hwang*, Youngjin Park and Younsik Park * Center for Noise and Vibration Control, Dept. of Mech. Eng., KAIST, Deajeon, Korea (Tel: +8-4-869-36, Email: tjdahr78@kaist.ac.kr) Abstract: We propose a sound source localization method using the Head-Related-Transfer-Function (HRTF) to be implemented in a robot platform. In conventional localization methods, the location of a sound source is estimated from the time delays of wave fronts arriving in each microphone standing in an array formation in free-field. In case of a human head this corresponds to Interaural-Time-Delay (ITD) which is simply the time delay of incoming sound waves between the two ears. Although ITD is an excellent sound cue in stimulating a lateral perception on the horizontal plane, confusion is often raised when tracking the sound location from ITD alone because each sound source and its mirror image about the interaural axis share the same ITD. On the other hand, HRTFs associated with a dummy head microphone system or a robot platform with several microphones contain not only the information regarding proper time delays but also phase and magnitude distortions due to diffraction and scattering by the shading object such as the head and body of the platform. As a result, a set of HRTFs for any given platform provides a substantial amount of information as to the whereabouts of the source once proper analysis can be performed. In this study, we introduce new phase and magnitude criteria to be satisfied by a set of output signals from the microphones in order to find the sound source location in accordance with the HRTF database empirically obtained in an anechoic chamber with the given platform. The suggested method is verified through an experiment in a household environment and compared against the conventional method in performance. Keywords: Sound source localization, Head-Related-Transfer-Function (HRTF), Phase criterion, Magnitude criterion 1. Introduction The sound source localization is about finding the whereabouts of a sound source using measurements from a number of microphones. The studies for developing localization model have a long history and many researchers have studied different methods for sound source localization. These days, mobile robot technology is gaining much attention in many application fields. The sound localizing ability of a robot is essential for human-robot communication and interaction. A robot operating in a household environment should detect diverse sound events and take notice of them to achieve robust recognition and interaction with user. So, sound source localization can be said to be one of the cores of the robot technology. ITD (Interaural Time Delay) plays an important role in most conventional methods for localization. Although many different sound source localization methods such as beamforming[1], spatial spectrum[], biological cues are developed, the ITD method[3] is one of the most popular methods in practical applications. ITD indicates the time delay between two microphones when acoustic waves emitted from a sound source reach each microphone. The ITD method estimates time delay and localizes the sound source with free-field assumption. Although ITD is an excellent sound cue in stimulating a lateral perception on the horizontal plane, confusion is raised when estimating the sound source location from ITD alone because many positions sharing the same ITD in 3-dimensional space can exist[4]. The ITD method also assumes that microphones are placed in free-filed, but this assumption is not valid for an actual platform used in real environments. For example, microphones are embedded in the robot head. Therefore the phase and magnitude of signals are distorted due to diffraction and scattering by the shading object such as the head and body of the platform. HRTF (Head-Related-Transfer-Function) summarizes the direction dependent acoustic filtering which a free-filed sound undergoes due to the head, torso, shoulder and pinna[]. HRTF associated with a dummy head microphone system or a robot platform with several microphones contains not only the information regarding proper time delays but also phase and magnitude distortions. So, we propose a new localization method using HRTF database empirically obtained in an anechoic chamber with a given platform. Performance of the proposed method is shown through experiments carried out in an anechoic chamber and a household environment. In addition, appropriate filtering method of noise existing in daily environment is proposed and the result is shown.. HRTFs for Dummy head In this paper, we apply the proposed method to the B&K HATS. First, we took measurements and constructed the HRTF database of the dummy head with azimuth varying from to 18 and elevation from -3 to 9 in an anechoic chamber. The sampling frequency was 44.1k. The HRTFs were calculated by dividing the pressure at each ear by the free-field pressure at the center of the head. Figure 1 shows the HRTFs for horizontal plane sources from m. When the source is placed at the very front and back of the head, the left and right ear HRTFs are the same due to symmetry of the head. As the source moves in a counterclockwise direction, the magnitude of the left ear HRTF increases and that of the right ear decreases due to the shadowing effect. However, when the source is located right in front of the left ear, the so-called bright spot occurs at the right ear: All the waves propagating around the head arrive at the right ear in phase resulting in a slight magnitude boost. As HRTFs vary according to change of azimuth, HRTFs also vary with change of elevation. Figure shows magnitude of the left ear HRTF for median plane (azimuth= ) sources from 1m. The main causes of variety are diffraction, reflection, and scattering by the torso, shoulder, and pinna.

ICCAS 3 1-1 - -3-4 - 3 1-1 - -3-4 1 1 3 1 4 frequency () - 1 1 3 1 4 3 frequency () 6 9 3 1-1 - -3-4 - 3 1 1 3 1 4 frequency () 3 3 1-1 - -3-4 - 3 left HRTF right HRTF 1 1 3 1 4 frequency () June -, KINTEX, Gyeonggi-Do, Korea If the noise can be ignored, the phase difference and magnitude ratio between the two ear outputs can be directly obtained from the HRTFs, corresponding to the actual location of sound source. So, we can detect the sound direction by finding the θ HRTF, M HRTF set minimizing the phase and magnitude criteria and this set directly corresponds to the actual location of the source. Coherence function can be used a weighting function. It is a measure of evaluating linear relationship between the two signals and represents how much uncorrelated noise contaminates the signals. As a result, we can reduce the uncorrelated noise effects by using the coherence function as a weighting function. 1-1 - -3-4 - 1-1 - -3-3 -4-4 - - 1 1 18 1 1 1 3 4 1 1 3 1 4 1 1 3 1 4 frequency () frequency () frequency () Fig. 1 Magnitude of HRTF for horizontal plane 1 1 - -1-1 - 6 - -3 1 1 1 1 3 1 4 frequency () 1 1 - -1-1 - - 1-1 - 9-3 1 1 3 1 4 frequency () 4. Experiment in an anechoic chamber 4.1 Azimuth estimation Figure 3 and Figure 4 show the calculated phase and magnitude criteria on horizontal plane with varying azimuth of an actual sound source in an anechoic chamber and the voice frequency band (VFB, i.e. 3 ~ 4 ) is used for calculation. For comparison with the ITD method, the phase criterion using free-field data, calculated from eq. (1) with replacing θ by HRTF θ ff, is also shown. θ ff is the phase difference between the two ears under the free-field assumption and it can be analytically calculated as follows. - -1-1 - - 3 θ ff ( f ) = π fτ, τ : ITD (3) -3 1 1 3 1 4 frequency () 1 1 - -1-1 - - -3 1 1 1 1 3 1 4 frequency () - -1-1 - - 3-3 1 1 3 1 4 frequency () Fig. Magnitude of left ear HRTF for median plane 3. Localization Cues A set of HRTFs for any given platform provides a substantial amount of information about whereabouts of the sound source. In this study, we introduce new phase and magnitude criteria to be satisfied by a set of output signals from a dummy head microphone system in order to find the sound source location in accordance with the HRTF database empirically obtained in the anechoic chamber. The phase and magnitude criteria are defined as follows. e { ( ) ( )} = γ θ ω θ ω d ω (1) ω phase HRTF { ( ) ( )} e = γ M ω M ω d ω () ω mag HRTF γ : coherence between right and left ear outputs θ : phase differnece between right ear and left ear outputs M : magnitude ratio between right ear and left ear outputs θhrtf : phase differnece between right ear HRTF and left ear HRTF M HRTF: magnitude ratio between right ear HRTF and left ear HRTF Fig. 3 Phase criterion for azimuth estimation Fig. 4 Magnitude criterion for azimuth estimation From Figure 3, it can be seen that the phase criterion calculated from HRTFs corresponding to the actual source location has a minimum value at the true angle. However, there is another HRTF angle, which is almost symmetric

ICCAS position about the interaural axis, making the phase criterion low. This means that the front-back confusion results from the phase assessment alone and this confusion also appears in the free-field result. In general, the estimation performance using the HRTF is better than that using the free-field data. Specially, as the source leans toward one ear, the scattering and diffraction effects for the hidden ear due to the head are most dominant. As a result, the phase criterion is faint in the free-field result. On the other hand, it is clear in the result using the HRTF data because HRTFs contain the information about the scattering and diffraction due to the shading object. In Figure 4, the magnitude criterion doesn t give sufficient information about the source location. Although low values of criterion are faintly shown, we can not determine the azimuth of the actual source. From the results, it can be said that the estimation performance for azimuth localization based on the HRTFs is better than the performance under the free-field assumption. And the phase criterion is more useful for azimuth localization than the magnitude criterion. 4. Elevation estimation Figure shows the phase and magnitude criteria calculated in the voice frequency band with varying elevation of a sound source from -3 to 3 at some selected azimuth. According to the result of the phase criterion, the criterion has low value not only at the HRTF angle corresponding to the true angle but also at the symmetric position to that. It can be said that up-down confusion is generated as the front-back confusion occurred in the azimuth localization case. However, the shape is somewhat different from that of azimuth due to the vertical asymmetry of the dummy head about the interaural axis. The both sides about the median plane are almost symmetric whereas, the upper and lower halves of the dummy head are asymmetric about the horizontal plane. The magnitude criterion contains this asymmetry, therefore the magnitude criterion has the minimum value at the HRTF angle corresponding to the true angle without confusion in estimation. As a result, for elevation estimation, it can be said that the magnitude criterion is more useful than the phase criterion. June -, KINTEX, Gyeonggi-Do, Korea. Experiment in a household environment Earlier works are conducted in an anechoic chamber and this means that we have no regard for noise. However, in daily environment, many noise sources such as the ambient noise, reflection, reverberation exist. Here, the proposed method is verified in a household environment and the method to reduce the noise effects is introduced in the following..1 Azimuth estimation Figure 6 shows the experimental results. For verification of the proposed method, the result based on the conventional method which uses ITD calculated by GCC (Generalized Cross-Correlation method) with the free-field assumption is shown. In the range of azimuth from about 6 to 9, the ITD in the free-field exceeds the imum value, ITD. When the distance between the two ears is a, ITD is determined by a ITD =, c : speed of sound (4) c As a result, the ITD method fails to estimate an accurate sound source location and this arises due to the noise contaminating the cross-correlation between the two ear outputs. On the other hand, the proposed method based on the phase criterion can detect the azimuth in general without noise filtering. However, there is an error in estimation. estimated angle (degree) 1 9 8 7 6 4 3 The results using the phase criteria and ITD based on GCC the phase criteria ITD based on GCC 1 (a) (b) Fig. Phase and magnitude criteria (a) azimuth = 3, (b) azimuth = 6 1 3 4 6 7 8 9 actual angle (degree) Fig. 6 Azimuth estimation results. Elevation estimation Table 1 shows the experimental result for elevation localization at some selected azimuth. The estimation performance is poor because the magnitude ratio between the two ear outputs is contaminated by background noise, reflection from household goods and secondary sound sources. An example of this contaminating effect is shown in Figure 7. Ideally, or in the anechoic chamber, the magnitude ratio between the two ear outputs should match one of the magnitude ratios between the two ear HRTFs and this HRTF set corresponds to the sound source location. However, if noise exists, the information about the magnitude of pure output signals becomes inaccurate.

ICCAS Table 1 Elevation estimation results Elevation Azimuth (degree) (degree) 3 6 9-1 -1 1-1 -1 1 1-1 1-1 1 June -, KINTEX, Gyeonggi-Do, Korea this incompleteness arises from the noise, thus the absolute value of phase in a household environment is not in accordance with that in the anechoic chamber although the group delays, which mean the gradient of phase, are almost the same except at several frequencies that noise seriously distorts. However, completely unwrapped phase can be obtained by the filtering and the phase of closely follows the anechoic chamber data. HRIR Filter Fig. 7 An example of contaminated magnitude ratio HL(azimuth deg) Fig. 8 HRIR filtering HR(azimuth deg) Up to the present, the experimental results of the proposed method without noise filtering are shown. Although the estimation performance in an anechoic chamber is good, the performance is poor in a household environment due to noise effects. As a result, for precise localization in the real world, noise reduction is necessary and this has direct relation to the estimation performance. 1 1-4 6 8 1 1 14 16 18 - -4-6 -8..4.6.8 1 1. 1.4 1.6 1.8 x 1 4 1 1-1 4 6 8 1 1 14 16 18 - -4-6 4 6 8 1 1 14 16 18.3 Filtering of HRIR For noise reduction, we propose a filtering of HRIR (Head Related Impulse response). HRIR is a time domain version of HRTF and it can be obtained by the inverse Fourier transform of HRTF. Figure 8 shows an example of measured HRIR in a household environment. In this figure, the small ripples representing the background noise in the room and reflections from household furniture can be observed by the irregular shape of the second peak, third peak and so on. However, the part related with the first peak is almost uncontaminated to background noise or reflections, so it can be said that this part directly reflects the pure effect by the actual sound source. Thus we can get rid of the noise effect and obtain uncontaminated HRIRs by applying a filtering as shown in Figure 8 and the length, having the unity value as its magnitude, corresponds to the meaningful length of the first peak part in the measured HRIR in an anechoic chamber. By taking the fast Fourier transform to this filtered HRIR again, we can get a which is almost noise-free. In Figure 9, some s are shown with HRTFs in a household environment and an anechoic chamber. Through the filtering, small ripples in the magnitude of HRTF are smoothed out, thus the almost agrees with that of the anechoic chamber. In addition, the phase of filtered HRTF is almost the same with that from the anechoic chamber and the filtering overcomes the problem that experimental HRTF phase is different from that of the anechoic chamber s HRTF due to incompleteness of unwrap. It can be said that HL(azimuth 3deg) 1 1 1 4 6 8 1 1 14 16 18 - -4-6 4 6 8 1 1 14 16 18 HL(azimuth 6deg) 1 4 6 8 1 1 14 16 18 - -4-6 4 6 8 1 1 14 16 18 HL(azimuth 9deg) 1 4 6 8 1 1 14 16 18 - -4-6 4 6 8 1 1 14 16 18 HR(azimuth 3deg) 1 1-4 6 8 1 1 14 16 18 - -4-6 -8..4.6.8 1 1. 1.4 1.6 1.8 x 1 4 HR(azimuth 6deg) 1 1-4 6 8 1 1 14 16 18 - -1..4.6.8 1 1. 1.4 1.6 1.8 x 1 4 HR(azimuth 9deg) 1 1-4 6 8 1 1 14 16 18 - -4-6 -8 4 6 8 1 1 14 16 18 Fig. 9 Magnitude and phase of HRTFs

ICCAS.4 Localization using the s By filtering the HRIR, we can obtain the almost uncontaminated to noise and apply the proposed localization algorithm using the phase and magnitude criteria based on the. Figure 1 shows the experimental result for azimuth localization on the horizontal plane. As mentioned before, azimuth is estimated based on the phase criterion. estimated angle (degree) 9 8 7 6 4 3 1 azimuth estimation 1 3 4 6 7 8 9 actual angle (degree) Fig. 1 Azimuth estimation results using the From the result, it is clear that the proposed method using the has the ability of precise azimuth estimation. However, since the HRTF database is obtained on the horizontal plane from to 18 with 1 increment, the above result does not indicate that we can find the source location within an accuracy of several degree. That is to say, we can localize the sound source within an accuracy of about 1 and if we construct the HRTF database with less increment, the resolution of estimation can be increased. The result of elevation experiment is shown in Table. Elevation testing was performed for -1,, 1, with the azimuth sets of 3, 6, 9, respectively. For localization, the magnitude criterion is used. Table Elevation estimation results using the Elevation Azimuth (degree) (degree) 3 6 9-1 -1-1 -1 1 1 1 1 1 June -, KINTEX, Gyeonggi-Do, Korea 6. Conclusion In this paper, we describe a sound source localization method using HRTF database. The phase difference and magnitude ratio between the two microphones are good localization cues and the HRTF contains information about that. Based on this, we propose two localization cues which are the phase and magnitude criteria and show experimental results using these cues in an anechoic chamber. Experimental results in a household environment are also shown. Although the estimation performance in the anechoic chamber is good, the performance is poor in the household environment due to the noise effects such as reflection, background noise, and additional sources. For reducing the noise effects, we propose a filtering of HRIR and this yields the. Although the filter structure is simple, by using this filtering we can get appropriate. Based on these, we apply the proposed localization algorithm in the household environment and the estimation performance is improved. When using only two microphones, the conventional method cannot find the azimuth and elevation simultaneously, however, the proposed method which uses the HRTF database can overcome this problem. In the proposed method, we should know information about the free-field pressure since HRTF means the ratio of the surface pressure to the free-field pressure. However, in practical application, measuring the free-field pressure is not easy, so we will deal with the method which can localize the sound source without the information about the free-field pressure. This is left for future work. REFERENCES [1] M. Wax & T. Kailath, Optimum localization of multiple sources by passive array, IEEE Tran. On Acoustics, Speech and Signal Processing, vol. 31, no., pp. 11-117, Oct, 1983. [] R. Schmitdt, A signal subspace approach to multiple emitter location and spectral estimation, Ph. D thesis, Stanford University, Stanford, CA, USA, MUSIC, 1981. [3] M. S. Brandstein & H. F. Silverman, A robust method for speech signal time-delay estimation in reverberation rooms, Proc. ICASSP-97, vol. 1, pp. 37-378, April, 1997. [4] C. I. Cheng & G. H. Wakefield, Introduction to Head-Related transfer Functions (HRTFs): Representations of HRTFs in Time, Frequency, and Space, Journal of the Audio Engineering Society, vol. 49, no. 4, pp.31-48, 1. [] R. O. Duda & W. L. Martens, Range dependence of the response of a spherical head model,, Journal of Acoustic Society of America, 14 (), November, 1998. When comparing with the Table 1, which represents the results of elevation estimation without the noise filtering, the estimation performance is improved. Although about 1 estimation error exists, we can estimate the elevation of the sound source approximately and distinguish the ups and downs of the source position. For your information, conventional methods such as the ITD method cannot find the azimuth and elevation of a sound source simultaneously by using two microphones. Above result, however, shows that the proposed method can find both azimuth and elevation by using only two microphones.