A MICROPHONE ARRAY INTERFACE FOR REAL-TIME INTERACTIVE MUSIC PERFORMANCE

Similar documents
SOUND SPATIALIZATION CONTROL BY MEANS OF ACOUSTIC SOURCE LOCALIZATION SYSTEM

arxiv: v1 [cs.sd] 4 Dec 2018

Automotive three-microphone voice activity detector and noise-canceller

Exploiting a Geometrically Sampled Grid in the SRP-PHAT for Localization Improvement and Power Response Sensitivity Analysis

EXPERIMENTAL EVALUATION OF MODIFIED PHASE TRANSFORM FOR SOUND SOURCE DETECTION

A microphone array approach for browsable soundscapes

Airo Interantional Research Journal September, 2013 Volume II, ISSN:

Robust Low-Resource Sound Localization in Correlated Noise

Localization of underwater moving sound source based on time delay estimation using hydrophone array

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment

Michael Brandstein Darren Ward (Eds.) Microphone Arrays. Signal Processing Techniques and Applications. With 149 Figures. Springer

Auditory System For a Mobile Robot

Speaker Localization in Noisy Environments Using Steered Response Voice Power

Reducing comb filtering on different musical instruments using time delay estimation

A Fast and Accurate Sound Source Localization Method Using the Optimal Combination of SRP and TDOA Methodologies

Mel Spectrum Analysis of Speech Recognition using Single Microphone

BEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR

Subband Analysis of Time Delay Estimation in STFT Domain

Applying the Filtered Back-Projection Method to Extract Signal at Specific Position

Omnidirectional Sound Source Tracking Based on Sequential Updating Histogram

Direction-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method

Implementation of decentralized active control of power transformer noise

Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor. Presented by Amir Kiperwas

Joint Position-Pitch Decomposition for Multi-Speaker Tracking

Improving Meetings with Microphone Array Algorithms. Ivan Tashev Microsoft Research

Advances in Direction-of-Arrival Estimation

Multiple Sound Sources Localization Using Energetic Analysis Method

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.

TARGET SPEECH EXTRACTION IN COCKTAIL PARTY BY COMBINING BEAMFORMING AND BLIND SOURCE SEPARATION

ONE of the most common and robust beamforming algorithms

Sound Source Localization using HRTF database

Robust Speaker Identification for Meetings: UPC CLEAR 07 Meeting Room Evaluation System

Antennas and Propagation. Chapter 5c: Array Signal Processing and Parametric Estimation Techniques

Lab S-3: Beamforming with Phasors. N r k. is the time shift applied to r k

Simultaneous Recognition of Speech Commands by a Robot using a Small Microphone Array

Recent Advances in Acoustic Signal Extraction and Dereverberation

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming

WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY

A FAST CUMULATIVE STEERED RESPONSE POWER FOR MULTIPLE SPEAKER DETECTION AND LOCALIZATION. Youssef Oualil, Friedrich Faubel, Dietrich Klakow

ROBUST PITCH TRACKING USING LINEAR REGRESSION OF THE PHASE

Auditory modelling for speech processing in the perceptual domain

STAP approach for DOA estimation using microphone arrays

ENHANCED PRECISION IN SOURCE LOCALIZATION BY USING 3D-INTENSITY ARRAY MODULE

Broadband Microphone Arrays for Speech Acquisition

Figure 1. SIG ACAM 100 and OptiNav BeamformX at InterNoise 2015.

Separation and Recognition of multiple sound source using Pulsed Neuron Model

ACOUSTIC SOURCE LOCALIZATION IN HOME ENVIRONMENTS - THE EFFECT OF MICROPHONE ARRAY GEOMETRY

S. Ejaz and M. A. Shafiq Faculty of Electronic Engineering Ghulam Ishaq Khan Institute of Engineering Sciences and Technology Topi, N.W.F.

Dynamics and Periodicity Based Multirate Fast Transient-Sound Detection

A robust dual-microphone speech source localization algorithm for reverberant environments

Real time noise-speech discrimination in time domain for speech recognition application

Calibration of Microphone Arrays for Improved Speech Recognition

Speech Synthesis using Mel-Cepstral Coefficient Feature

Underwater Wideband Source Localization Using the Interference Pattern Matching

SOUND SOURCE LOCATION METHOD

ADAPTIVE ANTENNAS. TYPES OF BEAMFORMING

Electronically Steerable planer Phased Array Antenna

Mutual Coupling Estimation for GPS Antenna Arrays in the Presence of Multipath

LOCALIZATION AND IDENTIFICATION OF PERSONS AND AMBIENT NOISE SOURCES VIA ACOUSTIC SCENE ANALYSIS

PAPER Adaptive Microphone Array System with Two-Stage Adaptation Mode Controller

Proceedings of the 5th WSEAS Int. Conf. on SIGNAL, SPEECH and IMAGE PROCESSING, Corfu, Greece, August 17-19, 2005 (pp17-21)

AVAL AUDIO-VISUAL ACTIVE LOCATOR. Faculty Sponsor: Professor Kathleen E. Wage Kelly Byrnes Rony Alaghbar Jacob Cohen

Smart antenna for doa using music and esprit

Audio Restoration Based on DSP Tools

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter

A Low-Power Broad-Bandwidth Noise Cancellation VLSI Circuit Design for In-Ear Headphones

Different Approaches of Spectral Subtraction Method for Speech Enhancement

Comparison of LMS Adaptive Beamforming Techniques in Microphone Arrays

Lecture 3 Complex Exponential Signals

Fundamental frequency estimation of speech signals using MUSIC algorithm

SUPERVISED SIGNAL PROCESSING FOR SEPARATION AND INDEPENDENT GAIN CONTROL OF DIFFERENT PERCUSSION INSTRUMENTS USING A LIMITED NUMBER OF MICROPHONES

ROOM SHAPE AND SIZE ESTIMATION USING DIRECTIONAL IMPULSE RESPONSE MEASUREMENTS

ACOUSTIC DATA TRANSMISSION IN AIR USING TRANSDUCER ARRAY

Acoustic Source Tracking in Reverberant Environment Using Regional Steered Response Power Measurement

Improving room acoustics at low frequencies with multiple loudspeakers and time based room correction

Proceedings of Meetings on Acoustics

Acoustic Change Detection Using Sources of Opportunity

Acoustic signal processing via neural network towards motion capture systems

Acoustic Beamforming for Hearing Aids Using Multi Microphone Array by Designing Graphical User Interface

NOISE ESTIMATION IN A SINGLE CHANNEL

(i) Understanding the basic concepts of signal modeling, correlation, maximum likelihood estimation, least squares and iterative numerical methods

Source Localisation Mapping using Weighted Interaural Cross-Correlation

Using Delay Estimation to Reduce Comb Filtering of Arbitrary Musical Sources

Multiple sound source localization using gammatone auditory filtering and direct sound componence detection

K.NARSING RAO(08R31A0425) DEPT OF ELECTRONICS & COMMUNICATION ENGINEERING (NOVH).

Location of Remote Harmonics in a Power System Using SVD *

Lab 4 Digital Scope and Spectrum Analyzer

Ocean Ambient Noise Studies for Shallow and Deep Water Environments

Directivity Controllable Parametric Loudspeaker using Array Control System with High Speed 1-bit Signal Processing

EXPERIMENTS IN ACOUSTIC SOURCE LOCALIZATION USING SPARSE ARRAYS IN ADVERSE INDOORS ENVIRONMENTS

Active noise control at a moving virtual microphone using the SOTDF moving virtual sensing method

VIBRATO DETECTING ALGORITHM IN REAL TIME. Minhao Zhang, Xinzhao Liu. University of Rochester Department of Electrical and Computer Engineering

Testing of Objective Audio Quality Assessment Models on Archive Recordings Artifacts

Speech Enhancement using Wiener filtering

Microphone Array Design and Beamforming

DESIGN OF VOICE ALARM SYSTEMS FOR TRAFFIC TUNNELS: OPTIMISATION OF SPEECH INTELLIGIBILITY

Performance Analysis of MUSIC and MVDR DOA Estimation Algorithm

29th TONMEISTERTAGUNG VDT INTERNATIONAL CONVENTION, November 2016

Audio Fingerprinting using Fractional Fourier Transform

Reverberant Sound Localization with a Robot Head Based on Direct-Path Relative Transfer Function

Transcription:

A MICROPHONE ARRA INTERFACE FOR REAL-TIME INTERACTIVE MUSIC PERFORMANCE Daniele Salvati AVIRES lab Dep. of Mathematics and Computer Science, University of Udine, Italy daniele.salvati@uniud.it Sergio Canazza Sound and Music Computing Group Dep. of Information Engineering, University of Padova, Italy canazza@dei.unipd.it Gian Luca Foresti AVIRES lab Dep. of Mathematics and Computer Science, University of Udine, Italy gianluca.foresti@uniud.it ABSTRACT A novel digital musical interface based on sound source localization using a microphone array is presented. It allows a performer to plan and conduct the expressivity of a performance by controlling an audio processing module in real-time through the spatial movement of a sound source (i.e., voice, traditional musical instruments, and sounding mobile devices). The prototype interface consists of an adaptive parameterized Steered Response Power Phase Transform (SRP-PHAT) with a Zero-Crossing Rate (ZCR) threshold and a Kalman filter that provides a more accurate estimate and tracking of the source position if there is movement. A real-time software based on external Max object was developed to test the system in a real-world moderately reverberant and noisy environment, focusing on the performance of pseudo-periodic sounds in a multisource scenario. 1. INTRODUCTION Recently, microphone array signal processing is increasingly being used in human computer interaction systems, for example the new popular interface Microsoft Kinect incorporates a microphone array to improve the voice recognition using the acoustic source localization and the beamforming for noise suppression. In the past years, a large number of musical interfaces has been implemented with the goal of providing tools for gestural interaction with digital sounds, using systems played by touching or holding the instrument, interfaces with haptic feedback, systems worn on the body, and interfaces that may be played without any physical contact (electric field sensors [12], optical sensors [7], ultrasound systems [10], and video camera that allows the performers to use their full-body for controlling in real-time the generation of an expressive audio-visual feedback [1]). This paper presents a novel digital musical interface for real-time interactive music performance, which uses a microphone array to estimate the sound source position in the plane and to allow a performer to use the two x- y coordinates of position to control an audio processing module in real-time through the spatial movement of a sound source. Musical interfaces are often used to allow the performer to enhance the expressive control on the sounds generated by their acoustic instruments in a live electronics context. E.g., in the works by Adriano Guarnieri - Medea (02) and Fili bianco-velati (05) - produced at the Centro di Sonologia Computazionale of Padova, the movement of a musician is followed by a motion capture system based on infrared cameras to control a live electronics patch [4], and using the robust, but very expensive, PhaseSpace optical motion capture system. It is composed by led systems, video cameras, and calibration procedure. In general, those kind of systems have considerable complexity and in some situations there could be problems with the low and/or not always controllable lighting of the concert hall, even when using infrared camera. It has been shown in [14] that there is some potentiality in using the sound source localization to directly control the position of a sound played back through a spatialization system by moving the sound produced by its own musical instrument. This work has been improved in [13] introducing an adaptive parameterized Generalizated Cross-Correlation (GCC) PHAT filter to localize musical sounds that are mainly harmonics. Both interfaces [14] [13] are been tested in a controlled real environment without verifying how the system works with interfering sources from a sound reinforcement system and other instruments. Thus, in this paper a validation in multi-source scenario is presented, introducing the adaptive parameterized SRP-PHAT with a ZCR threshold (Section 3) that has a better performance than the parameterized GCC-PHAT proposed in [13] as shown in Section 4. _473

ZCR estimation threshold µ x1(t) sound acquisition x2(t) x3(t) maximum peak detection xy estimation Kalman filter xy mapping control audio processing audio processing Figure 1. Block diagram of interface. 2. SSTEM ARCHITECTURE The proposed interface has the advantage of being completely non-invasive (no need for markers, sensors or wires on the performance), and requires no dedicated hardware. The architecture consists of a microphone array and digital signal processing algorithms for robust sound localization. Figure 1 summarizes the system architecture of the interface. The array system is composed of three halfsupercardioid polar pattern microphones, which reduce ambient noise and pickup of room reverberation, arranged in an uniform linear placement. In this way, we can localize a sound source in a plane (three microphones are the bare minimum). Signal processing algorithms estimate the sound source position in an horizontal plane by providing its Cartesian coordinates. A SRP-PHAT method is used to compute the acoustic map analysis. To improve the performance in the case of harmonic sounds, or generally pseudo-periodic sounds, a parameterized SRP-PHAT is proposed. The ZCR function is used to determinate if a sound is pseudo-periodic, and to adapt with a threshold value the parametrically control of PHAT filter. A Kalman filter [8] is applied to smooth the time series of observed position to obtain a more robust and accurate xy values. The last component implements the mapping strategy [16] to associate the x-y coordinates with audio processing parameters. 3. ADAPTIVE PARAMETERIZED SRP-PHAT WITH ZERO-CROSSING RATE THRESHOLD The SRP-PHAT [5] is based on the concept of adding several time delay estimation functions from the microphone pairs. It consists of calculating the GCC-PHAT function between pairs of microphones and using the Global Coherence Field (GCF) [11] to construct a spatial analysis map to improve the localization performance. Given the vector s =[xy] T of space and R microphone pairs, the SRP-PHAT at time t can be expressed S (s, f [R GCC (t)]) = R r=1 f [R GCC r (t)] (1) where the R GCC (t) is the GCC-PHAT of the r th pair. The position of the source is estimated by picking the maximum peak ŝ n = argmaxs (s, f [R GCC (t)]). (2) s The GCC-PHAT [9] is the classic method to estimate the relative time delay associated with acoustic signals received by a pair of microphones in a moderately reverberant and noisy environment. The GCC-PHAT basically consists of a cross-correlation followed by a filter that aims to reduce the performance degradation caused by additive noise and multi-path channel effects. The GCC- PHAT in the frequency domain is R GCC x i x j (t)= 1 L 1 L f =0 j2π ft Ψ( f )S xi x j ( f )e L (3) where f is the Discrete Fourier Transform (DFT) integer frequency index, L is the number of samples of the observation time, and Ψ( f ) is the frequency domain PHAT weighting function, and the cross-spectrum of the two signals is defined as S xi x j ( f )=E{ i ( f ) j ( f )} (4) where i ( f ) and j ( f ) are the DFT of the signals x i (t) and x j (t) respectively, and * denotes the complex conjugate. GCC is used for minimizing the influence of moderate uncorrelated noise and moderate multipath interference, maximizing the peak in correspondence of the time delay. The PHAT weighting function places equal importance on each frequency by dividing the spectrum by its magnitude. The PHAT normalizes the amplitude of the spectral density of the two signals and uses only the phase information to compute the GCC Ψ PHAT ( f )= 1 S xi x j ( f ). (5) We note that SRP-PHAT, which uses the sum of the GCCs of the microphone pairs, is equivalent to using a steered response filter and sum beamforming with PHAT weighting. In fact, the SRP of a 2-element array is equivalent to the GCC of those two microphones. The SRP-PHAT algorithm has been shown to be one of the most robust sound source localization approaches operating in noisy and reverberant environments [15]. This algorithm enhances the performance of localization with a network of large arrays. However, the computational cost of the method is very high. To reduce the processing time of search algorithms, improvements have been suggested [2] [3]. _474

0 m1(-25,0) m2(0,0) m3(25,0) 25 25 S(-,) E(,) 75 25 0 25 75 25 0 25 Figure 2. The sample space position of a square with cm sides considering three microphones with distance of 25 cm and TDOAs between microphones m 1 -m 2 and m 2 - m 3. Sampling rate: 44110 Hz; 900 Hz. C1(-,90) sound source (x,y) area of movement x cm area of analysis map 0x0 cm C2(,90) It is important to note that the PHAT performance is dramatically reduced in the case of harmonic sounds, or generally pseudo-periodic sounds. In fact, the PHAT has less capability to reduce the deleterious effects of noise and reverberation when it is applied to a pseudo-periodic sound. An accurate analysis of the PHAT performance for a broadband and narrowband signal can be found in [6]. The results of this work highlight the ability of the PHAT to enhance the detection performance for single or multiple targets in noisy and reverberant environments, when the signal covers most of the spectral range. Thus, to work with pseudo-periodic sounds the proposal is to use a parameterized SRP-PHAT that weighs the contribution of the PHAT filtering, depending on the threshold of the ZCR parameters. The PHAT weighting can be generalized to parametrically control the level of influence from the magnitude spectrum [6]. This transformation will be referred to as the PHAT-β and defined as 1 Ψ PHAT β ( f )= S xi x j ( f ) β (6) where β varies between 0 and 1. When β = 1, equation (6) becomes the conventional PHAT and the modulus of the Fourier transform becomes 1 for all frequencies; when β = 0, the PHAT has no effect on the original signal, and we have the cross-correlation function. Therefore, in the case of harmonic sounds, we can use an intermediate value of β so that we can detect the peak to estimate the time delay between signals, and can have a system, at least in part, that exploits the benefits of PHAT filtering to improve performance in moderately reverberant and noisy environments. The results of localization improvement in case of pseudo-periodic sounds using the PHAT-β are reported in [13]. To adapt the value of β, we use the ZCR to determinate if the sound source is periodic. ZCR is a very useful audio feature and is defined as the number of times that the audio waveform crosses the zero axis ZCR(t)= 1 2L L i=1 where sgn(x) is the sign function. sgn(x(t + i)) sgn(x(t + i 1)) (7) Figure 3. The analysis map area with microphones and interference sources (pink noise) position (C 1 and C 2 ). In the experiments the flute sound source moves from point S to point E. Then, we can express the adaptive parameterized GCC- PHAT, identifying by experimental tests a suitable threshold µ such as { β = 1, i f ZCR µ (8) β < 1, i f ZCR < µ 4. EPERIMENTAL RESULTS The experiments to verify the performance of interface in a multi-source scenario were conducted in a rectangular room of 3.8 4.4 m, in a moderately reverberant (RT=0.35 s) and noisy environment. The experiments were made with a real-time interface developing a Max external objects for interactive performance. The interface works with sampling rate of 96 khz, a Hann window analysis of 42 ms, and a distance between microphones of 25 cm. The sampling rate and the microphone distance determine the number of samples to estimate the TDOA. In fact, by increasing these parameters it is possible to obtain a higher resolution on the TDOA, and therefore a higher resolution in the sample plane position. Figure 2 compares the sample space position with different sampling rate, considering a square with cm sides, three microphones with distance of 25 cm, and the TDOAs between microphones m 1 -m 2 and m 2 -m 3. The working area is located in a square with 1 meter sides. The axis origin coincides with the position of microphone 2, the x axis can vary between - cm and cm, and the y axis can vary between 0 and cm. It is important to note that when we use a SRP-PHAT in multisource case, we have to consider a largest area of analysis, otherwise competing sources can cause a peak that would be seen by the system as a false source in SRP-PHAT map inside the one meter square area. Therefore, for the experiments we consider a square area of analysis with 2 meter _475

30 10 0 10 30 0 SIR=30 db 30 10 0 10 30 SIR=25 db SIR= db 30 10 0 10 30 30 10 0 10 30 Figure 4. The x-y position estimation by the system of a flute sound with SIR=35 db. The Kalman filter data is the black line and raw data are the dots. GCC-PHAT-β sides; the peak of GCF map outside the area of interest is automatically discarded from the system. A flute sound (continuous G5 note) was used as source of interest, and pink noise, played from two loudspeakers, was used as interference source. The flute sound was played by a mobile device to better control and measure the movement and position of the source. Figure 3 shows the position of microphones (m 1,m 2,m 3 ), and of loudspeaker (C 1,C 2 ). The mobile device was made to manually move from point S to point E during the tests. Five experiments were made considering different Signal to Interference Ratio (SIR) defined as SIR = P f lute+noise P inter f erence+noise (9) where P f lute+noise is the average power detected at the microphone m 2 of the flute sound played by the mobile device in the xy position (0,) cm in absence of interfering sources, and P inter f erence+noise is the average power detected at the microphone m 2 of the pink noise interference sources. The localization source estimation has been done comparing GCC-PHAT-β and (for comparison between PHAT and PHAT-β see [13]). Figure 4 shows the estimation position with SIR=35 db. Both algorithms are very accurate. Increasing the intensity of the interferences, we can note that, when the SIR decreases, the capability of sound localization tends to fail, in particular when the angle of sound incidence on the array increases, since we are using half-supercardioid polar pattern microphones; we can note this especially for the GCC-PHAT-β (Figure 5). Moreover, when the SIR is 25 db the GCC-PHAT-β has more errors, compared to the. Table 1 summarizes the values of Root Mean Square Error (RMSE) for the variable y and shows the maximum and minimum estimates of the variable x. The RMSE of the variable y is computed considering the Figure 5. The x-y Kalman position estimation by the system with different SIR. GCC-PHAT-β GCC-PHAT-β SIR RMSE(y) min(x) max(x) RMSE(y) min(x) max(x) 35 1.9 -.4.4 1.4 -.2.3 30 3.3-39.2 37.1 1.7-35.5 38.9 25 3.5-36.7 32.5 2.5-33.6 34.6 3.9-30.9 29.7 2.1-33.2 33.6 15 nd nd nd nd nd nd Table 1. The values of RMSE(y), maximum and minimum of x related to the experiments. mobile device reference position of cm from the microphone array. We can see a good localization of the source up to SIR= db with, but at 15 db of SIR the localization source estimation is not possible for both algorithms by the system. Finally, an experiment involving a real flute instrument with β=0.85 and SIR=35 db is shown in Figure 6. The post-processing Kalman filter is a crucial and fundamental step to obtain a more accurate value of position and to track the source movement. 5. CONCLUSIONS A digital musical interface based on sound localization using a microphone array allows a performer to directly interact with a computer by moving a sounding object, and to plan and conduct the expressivity of a performance by controlling an audio processing module. A novel algorithm based on adaptive was proposed to solve the problem of acoustic source localization when the sound is pseudo-periodic. By developing a real-time software in Max environment, we obtained some experimental results that show how the system is able to locate the source in a more accurate and robust way than the GCC- PHAT-β, when SIR db in a multi-source moderate reverberant (RT=0.35 s) and noisy room. _476

10 raw data Kalman filter data [4] A. de Götzen, Enhancing engagement in multimodality environments by sound movements in a virtual space, IEEE Multimedia, vol. 11, pp. 4 8, 06. 30 70 90 30 10 0 10 30 Figure 6. The x-y position estimation of a real flute instrument with and SIR=35 db. Besides, it is shown how the angular resolution degrades the localization performance, in particular for the GCC-PHAT-β, in correspondence with the increase of competitive sources; locate is impossible when the SIR is less than db. An increase in the number of microphones could improve the system performance, especially with an algorithm based on SRP, which tends to have a more robust performance with a large array. The interface has considerably less complexity than systems based on electric field, optical and video camera sensors. The latter are widely used, especially for great success of Kinect, but in general it is possible to have problems for the low or variable lighting condition during live electronics. One of the most robust way is to use a motion capture system (e.g., PhaseSpace), which is very complex and expensive. Hence, we believe that our approach based on microphone array is presented as a viable alternative. However, in the future, we plan to use and test the microphone array interface in live electronic performances. 6. REFERENCES [1] G. Castellano, R. Bresin, A. Camurri, and G. Volpe, Expressive control of music and visual media by full-body movement, in Proceedings of the International Conference on New Interfaces for Musical Expression, 07, pp. 390 391. [2]. Cho, D. ook, and S. C. H. Kim, Sound source localization for robot auditory systems, IEEE Transactions on Consumer Electronics, vol. 55, no. 3, pp. 1663 1668, 09. [3] M. Cobos, A. Marti, and J. J. Lopez, A modified srp-phat functional for robust real-time sound source localization with scalable spatial sampling, IEEE Signal Processing Letters, vol. 18, no. 1, pp. 71 74, 11. [5] J. H. DiBiase, H. F. Silverman, and M. S. Brandstein, Microphone Arrays: Signal Processing Techniques and Applications. Springer, 01, ch. Robust localization in reverberant rooms. [6] K. D. Donohue, J. Hannemann, and H. G. Dietz, Performance of phase transform for detecting sound sources with microphone arrays in reverberant and noisy environments, Signal Processing, vol. 87, no. 7, pp. 1677 1691, 07. [7] N. Griffith and M. Fernstrom, Litefoot - a floor space for recording dance and controlling media. in Proceedings of the International Computer Music Conference, 1998, pp. 475 481. [8] R. E. Kalman, A new approach to linear filtering and prediction problems, Journal of Basic Engineering, vol. 82, pp. 35 45, 19. [9] C. Knapp and G. Carter, The generalized correlation method for estimation of time delay, IEEE Transactions on Acoustics, Speech and Signal Processing, vol. 24, no. 4, pp. 3 327, 1976. [10] I. Mott and J. Sosnin, A new method for interactive sound spatialisation, in Proceedings of the International Computer Music Conference, 1996, pp. 169 172. [11] M. Omologo and P. S. R. DeMori, Spoken Dialogue with Computers. Academic Press, 1998, ch. Acoustic Transduction. [12] J. Paradiso and N. Gershenfeld, Musical applications of electric field sensing, Computer Music Journal, vol. 21(3), pp. 69 89, 1997. [13] D. Salvati, S. Canazza, and A. Rodà, A sound localization based interface for real-time control of audio processing, in Proceedings of the 14th International Conference on Digital Audio Effects, 11, pp. 177 184. [14], Sound spatialization control by means of acoustic source localization system, in Proceedings of the 8th Sound and Music Computing Conference, 11, pp. 284 289. [15] H. F. Silverman,. u, J. M. Sachar, and W. R. I. Patterson, Performance of real-time sourcelocation estimators for a large-aperture microphone array, IEEE Transactions on Speech and Audio Processing, vol. 13, no. 4, pp. 593 6, 05. [16] V. Verfaille, M. Wanderley, and P. Depalle, Mapping strategies for gestural and adaptive control of digital audio effects, Journal of New Music Research, vol. 35, pp. 71 93, 06. _477