Applying the Filtered Back-Projection Method to Extract Signal at Specific Position

Similar documents
Estimation of Sinusoidally Modulated Signal Parameters Based on the Inverse Radon Transform

Mel Spectrum Analysis of Speech Recognition using Single Microphone

Basic Signals and Systems

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming

Digital Video and Audio Processing. Winter term 2002/ 2003 Computer-based exercises

Localization of underwater moving sound source based on time delay estimation using hydrophone array

Broadband Microphone Arrays for Speech Acquisition

Non Unuiform Phased array Beamforming with Covariance Based Method

Adaptive Fingerprint Binarization by Frequency Domain Analysis

Lab S-3: Beamforming with Phasors. N r k. is the time shift applied to r k

Audio Imputation Using the Non-negative Hidden Markov Model

Single Channel Speaker Segregation using Sinusoidal Residual Modeling

Audio Fingerprinting using Fractional Fourier Transform

COMPUTATIONAL IMAGING. Berthold K.P. Horn

Improving Meetings with Microphone Array Algorithms. Ivan Tashev Microsoft Research

The Analysis of the Airplane Flutter on Low Band Television Broadcasting Signal

Robust Low-Resource Sound Localization in Correlated Noise

Live multi-track audio recording

WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY

Chapter 1. Electronics and Semiconductors

Interference in stimuli employed to assess masking by substitution. Bernt Christian Skottun. Ullevaalsalleen 4C Oslo. Norway

Introduction to signals and systems

About Doppler-Fizeau effect on radiated noise from a rotating source in cavitation tunnel

Omnidirectional Sound Source Tracking Based on Sequential Updating Histogram

Antennas and Propagation. Chapter 6b: Path Models Rayleigh, Rician Fading, MIMO

Study on Multi-tone Signals for Design and Testing of Linear Circuits and Systems

Room Impulse Response Modeling in the Sub-2kHz Band using 3-D Rectangular Digital Waveguide Mesh

Speech Enhancement Using Microphone Arrays

Lecture 3 Complex Exponential Signals

Chapter 3 Data Transmission COSC 3213 Summer 2003

UNIT Explain the radiation from two-wire. Ans: Radiation from Two wire

SPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN. Yu Wang and Mike Brookes

Functions of more than one variable

Lecture 2: SIGNALS. 1 st semester By: Elham Sunbu

arxiv: v1 [cs.sd] 4 Dec 2018

SPECTRAL COMBINING FOR MICROPHONE DIVERSITY SYSTEMS

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.

Automotive three-microphone voice activity detector and noise-canceller

COMMUNICATION SYSTEMS NCERT

University Ibn Tofail, B.P. 133, Kenitra, Morocco. University Moulay Ismail, B.P Meknes, Morocco

Terminology (1) Chapter 3. Terminology (3) Terminology (2) Transmitter Receiver Medium. Data Transmission. Direct link. Point-to-point.

Introduction to Digital Signal Processing (Discrete-time Signal Processing)

Sound Source Localization using HRTF database

COMP 546, Winter 2017 lecture 20 - sound 2

Discrete Fourier Transform (DFT)

Measuring impulse responses containing complete spatial information ABSTRACT

Different Approaches of Spectral Subtraction Method for Speech Enhancement

Laboratory Assignment 4. Fourier Sound Synthesis

Outline / Wireless Networks and Applications Lecture 3: Physical Layer Signals, Modulation, Multiplexing. Cartoon View 1 A Wave of Energy

A comparative study on main lobe and side lobe of frequency response curve for FIR Filter using Window Techniques

Structure of Speech. Physical acoustics Time-domain representation Frequency domain representation Sound shaping

Direction-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method

Source Localisation Mapping using Weighted Interaural Cross-Correlation

KULLIYYAH OF ENGINEERING

Part 1: Standing Waves - Measuring Wavelengths

Sampling and Reconstruction of Analog Signals

A MICROPHONE ARRAY INTERFACE FOR REAL-TIME INTERACTIVE MUSIC PERFORMANCE

IMPULSE RESPONSE MEASUREMENT WITH SINE SWEEPS AND AMPLITUDE MODULATION SCHEMES. Q. Meng, D. Sen, S. Wang and L. Hayes

Off-axis response of Compton and photoelectric polarimeters with a large field of view

1. Measure angle in degrees and radians 2. Find coterminal angles 3. Determine the arc length of a circle

Signal Processing. Naureen Ghani. December 9, 2017

Approaches for Angle of Arrival Estimation. Wenguang Mao

Performance Analysis of Parallel Acoustic Communication in OFDM-based System

Joint Position-Pitch Decomposition for Multi-Speaker Tracking

11/8/2007 Antenna Pattern notes 1/1

DESIGN AND APPLICATION OF DDS-CONTROLLED, CARDIOID LOUDSPEAKER ARRAYS

Image Simulator for One Dimensional Synthetic Aperture Microwave Radiometer

Electromagnetic Spectrum

Signal Characteristics

Name Date Class. Identify whether each function is periodic. If the function is periodic, give the period

Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor. Presented by Amir Kiperwas

The Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals

THE CITADEL THE MILITARY COLLEGE OF SOUTH CAROLINA. Department of Electrical and Computer Engineering. ELEC 423 Digital Signal Processing

Fault Location Technique for UHV Lines Using Wavelet Transform

Presentation Outline. Advisors: Dr. In Soo Ahn Dr. Thomas L. Stewart. Team Members: Luke Vercimak Karl Weyeneth. Karl. Luke

Influences of a Beam-Pipe Discontinuity on the Signals of a Nearby Beam Position Monitor (BPM)

ON SAMPLING ISSUES OF A VIRTUALLY ROTATING MIMO ANTENNA. Robert Bains, Ralf Müller

LOCAL MULTISCALE FREQUENCY AND BANDWIDTH ESTIMATION. Hans Knutsson Carl-Fredrik Westin Gösta Granlund

Fourier Transform. Any signal can be expressed as a linear combination of a bunch of sine gratings of different frequency Amplitude Phase

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm

Chapter 6: Periodic Functions

Microphone Array project in MSR: approach and results

WARPED FILTER DESIGN FOR THE BODY MODELING AND SOUND SYNTHESIS OF STRING INSTRUMENTS

Proceedings of Meetings on Acoustics

Digital Loudspeaker Arrays driven by 1-bit signals

A Frequency-Invariant Fixed Beamformer for Speech Enhancement

Lab course Analog Part of a State-of-the-Art Mobile Radio Receiver

Beamforming in Interference Networks for Uniform Linear Arrays

Multiple Sound Sources Localization Using Energetic Analysis Method

Transforms and Frequency Filtering

Terminology (1) Chapter 3. Terminology (3) Terminology (2) Transmitter Receiver Medium. Data Transmission. Simplex. Direct link.

ECE 185 ELECTRO-OPTIC MODULATION OF LIGHT

Fiber Optic Communication Systems. Unit-04: Theory of Light.

RECENTLY, there has been an increasing interest in noisy

Trigonometry. An Overview of Important Topics

EXPERIMENTAL EVALUATION OF MODIFIED PHASE TRANSFORM FOR SOUND SOURCE DETECTION

Mobile Radio Propagation Channel Models

Implementation of a Real-Time Rayleigh, Rician and AWGN Multipath Channel Emulator

Chapter 4 SPEECH ENHANCEMENT

Airo Interantional Research Journal September, 2013 Volume II, ISSN:

Transcription:

Applying the Filtered Back-Projection Method to Extract Signal at Specific Position 1 Chia-Ming Chang and Chun-Hao Peng Department of Computer Science and Engineering, Tatung University, Taipei, Taiwan cmchang@ttu.edu.tw, g9167@mail.ttu.edu.tw Abstract In this paper, a back-projection method of extracting sound signal from multiple sources using microphone array is proposed. It is assumed that the signals received by microphone may be treated as a kind of non-straight projection. Therefore, the filtered back-projection method could be used to reconstruct the sound signal. The typical method used to enhance signal come from the focus location in multi-sources environment is the delay-and-sum method, whhich is based on the delay and shift operations. The filtered back-projection method usually is used to reconstruct the profile in the Computerized Tomography (CT). The algorithm to apply filtered back-projection method to extract the signal located at specific position is dressed in the paper. Experiments are also presented to prove the fact that filtered back-projection method could be used to extract the location-specified signal from the multi-source environment. 1. Introduction Microphone array is applied in many occasions nowadays. For example, concert, conference, speech, recording studio, etc [1]. Typically, delay and sum method is used for microphone array to enhance the signal come from specific location [2, 3]. We propose a different method in which the same function based on the computerized tomography technique is used. Microphone array receives the mixed signal from all sources. The signal received by microphone can be treated as a special projection. The signal from specific location can be reconstructed using the filtered back-projection method. Then, the signal from the location we focus on is extracted by the system. The organization of this paper is as follows. In Section 2, the delat-and-sum method and filtered back-projection method are addressed in brief. The discrete version of sound projection slice theorem is stated briefly in Section 3. The problems about the implementation of backprojection method to extract interested signal is discussed in Section 4. in Section 5, conclusions are addressed. 2. Relative researches The typical method of enhancing signal received by microphone array is the delay-and-sum method. The signal comes from a specific position are received by microphone array. These received signals are different intensity and time delay. The differences are due to the difference in distance from the specified location to each microphone as shown in Fig. 1. Let L k represent the distance from source to the k-th microphone, and L min be the minimum distance of L k. The delay time and wave propagation at- Figure 1: Delay-and-sum method tenuation are obtained from the follow equations, D k = (L k L min )/v (sec) (1) A k = L 2 k/l 2 min (amplitude), (2) where v is the velocity of sound. Applying D k and A k to the following equation, we can enhance the signal from the specific location, S e (t), S e (t) = 1 N N A k S k (t + D k ), (3) k=1 where N is the numbers of microphone, and S k is the signal received by the k-th microphone. Computerized Tomography (CT) is a useful technique to reconstruct the profile image of an object without destroying it from the projections. A projection is a shadowgram obtained by illuminating an object by penetrating radiation. The Radon transform and its inverse were formulated by J. Radon in 1917 [4]. The Radon transform provides the mathematical framework necessary for forth and going back between the spatial coordinates {x, y} and the projection space coordinates {s, θ}. The Radon transform of a function f(x, y) at angle θ, denoted as g(s, θ), is defined as the line integral along a line inclined at an angle from the u-axis and at a distance s from the origin as shown in Fig. 2. The Radon transform is defined as g(s, θ) = R{f}

2 Figure 3: The Radon geometry Figure 2: The Radon geometry = f(x, y)δ(x cos θ + y sin θ s) dxdy, (4) where < s < and θ < π. The symbol R, denoting the Radon transform operator, is also called the projection operator. In the rotated coordinate system (s, u), the relationship to (x, y) are s = x cos θ + y sin θ (5) u = x sin θ + y cos θ. (6) Associated with the Radon transform is the backprojection operator denoted as β and is defined as b(x, y) = β{g(s, θ)} = g(x cos θ + y sin θ, θ) dθ. (7) The quantity b(x, y) is called the back projection of g(s, θ). In polar coordinates it can be written as b p (s, φ) = g(r cos(θ φ), θ) dθ (8) Eqs. (7) and (8) represent the accumulation of the raysums of all of the rays that pass through the point (x, y). For example, if g(s, θ) = g 1 (s)δ(θ θ 1 ) + g 2 (s)δ(θ θ 2 ) (9) that is, there are only two projections, then b p (r, φ) = g 1 (s 1 ) + g 2 (s 2 ), (1) where s 1 = r cos(θ 1 φ) and s 2 = r cos(θ 2 φ). In general, for a fixed point (x, y) or (r, φ), the value of back projection β{g} is evaluated by integrating g(s, θ) over θ for all lines pass through that point. The one-dimensional Fourier transform with respect to s of the projection g(s, θ) can be obtained from Eq. (4) G θ = F 1 {g θ (s)} = f(s cos θ u sin θ, s sin θ + u cos θ) e j2πrs dsdu (11) where F 1 is the one-dimensional forward Fourier transformation and G θ (r) is the Fourier transform of g θ (s). Rotating the coordinate from {s, u} to {x, y} then Eq. (11) becomes G θ (r) = f(x, y)e j2π(xr cos θ+yr sin θ) dxdy = F (r cos θ, r sin θ) (12) where F (r cos θ, r sin θ) is a slice of two-dimensional Fourier transform of tomogram at angle θ as shown in Fig. 3. If θ < π and < r <, F (r cos θ, r sin θ) represents the two-dimensional Fourier transform of the tomogram in polar form. The tomogram can be obtained by applying the two-dimensional inverse Fourier transform in polar form, that is, F (r cos θ, r sin θ) r e j2πr(x cos θ+y sin θ) drdθ (13) where dr represents the one-dimensional inverse Fourier transform, g dθ denotes the back projection as Eq. (8), and F (r cos θ, r sin θ) r represent the filter operation in spatial domain. Combining the equations above, we have F 1 1 {F 1{g θ (s)} r } dθ, (14) where g θ (s) is the projection of f(x, y) at angle θ, the F 1 and F1 1 are the one-dimensional Fourier and inverse Fourier transforms, respectively. And r is the backprojection filter in frequency domain and s = x cos θ + y sin θ. The slice information we have are discrete in angle. Then, the Eq. (14) in discrete form is F1 1 {F 1{g θ (s)} r }. (15) θ<π Therefore, profile of the object, that is f(x, y), could be reconstructed by summation of each filtered slice g θ (s). 3. Sound Projection We may consider that if the sound signal received by each microphone is a kind of projection, a non-straight projection, unlike the x-ray is doing. Then we can use the

3 Figure 4: The sound projection. back-projection technique to reconstruct the sound source signal. The signals come from point sources located on the concentric circules are received by microphones at the same time if the microphones is at the center. In other word, if the distances from sources to one microphone are the same, signal from different sources on the same concentric circule reach to the microphone at the same time. Fig. 4 shows the relation between distance from microphone and received time, where d k is the radius from center, the position of microphone, and t k is the time that wave propagated to microphone. The received signals at time T +t k is the summation on radius d k and is multiplied by a constant attenuation coefficient. We may consider that would exist a sound projection slice theorem, then the back-projection operation of sound could be rewrite from Eq. (15) and obtain s(t ) = k<m F 1 1 {F 1{m k (T )} ω }, (16) where T is a time period, s(t) is the source signal, m k (t) is the signal received by the kth microphone, ω is the back-projection filter in frequency domain, and M is the size of microphone array. Next, we use the Eq. (16) to separate the source we focus on from mixed-signal received by microphone array. 4. Implementation For the limitation of hardware, we use Matlab to simulate all of the system. The system structure is shown in Fig. 5. Assume there is a conference room with 4 meters width and 5 meters length without reverberant. A 16-channel microphone array is linearly arranged on the bottom side. Two point sources could be placed at any place in this room. The sine wave is used as the sound source in the begining of simulation. The main part of the program is the filtered backprojection operation. The sequence is choosing the proper region of signal received from each microphone and filtering it, then average the filtered signal of each microphone in time domain like delay and sum method. In Fig. 6, the simulation of filtered back-projection method with two sine wave signal sources is shown. The signals of two sources is the same as shown at the up-left plot in Fig. 6. The microphone receive the mixed-signal from source (the signal is interested) and noise (the signal isn t interested) is shown in bottom-left plot in Fig. 6. The Figure 5: System architecture. Figure 6: Simulation of filtered back-projection method with two sound sources. result in spacial domain is shown at bottom-right plot in Fig. 6. The equation used to measure the efficiency to extract the interesting signal is shown as ( ) s SNR signal = 2 log (db), (17) s r where s and r are the signals of the source and extraction result, respectively. This measurement shows the extracted signal compares with original signal. Another measurement is used to measure the capability to reduce noisy signal is shown as ( ) r s SNR noise = 2 log (db), (18) s where s and r are the signals of the source and extraction result, respectively, and s denotes the noisy signal. The simulation results of sound source extraction are presented. First, sources in different frequencies are sim-

4 Figure 7: Simulation of two sound sources with two differ- Figure 8: Simulation of two single frequency sources with ent frequency using filtered back-projection method with. signal fixed position and noise changed. The higher value ulated. Second, experimation with changing the microphone array size and arrangement are shown. Finally, two pieces of real music are mixed and separated. First, we verify the algorithm with two different frequencies in each source. The arrangement of components is the same as above. Let S1 and S2 are two dual frequency sources. The frequencies of S1 are 1135 and 341 Hz. The frequencies of S2 are 14 and 238 Hz. In the Fig. 7 the spectrums of all the signal are shown. The top two figures are the spectrum of S1 and S2. The middle two figures are the signal received by microphone, m8. The left side figure: S1 is treated as the signal that we want then S2 is noise. On the contrary, the right side figure S2 is treated as signal. In the result, the magnitude of low frequency is smaller then high frequency. This is due to the filter we used is the high pass filter. Next, we verify the efficiency and the sources position relation in two scenarios. Two point sources are placed in the conference room. One is treated as signal and the other is treated as interference. The signal source is placed at the fixed position and interference source is moved in the room area. When noise located at each point, SNRsignal and SNRnoise are calculated. The moving step of interference is 1 cm. In order to show the tendency of extraction efficiency, room size is changed to 6 6 meters and microphone array is still placed at the same place with 4 meters width. The single frequency and multi-frequency results are shown in Figs. 8 and 9. The second scenario, we let the interference at a fixed position and move source in the room. Then also calculate the SNR with the same condition. Figs. 1 and 11 are the result of the single and multiple frequencies, respectively. From the four figures above, we could see the dark part means the simulation could extract signal we want clearly when signal source is in front of the noise source and near the microphone array. On the contrary, when signal is near the noise or far from to microphone array, the signal to noise ratio is less than 2 db. At final, we use two pieces of music in the simulation. The S1 is female voice song and S2 is the male voice song. The simulation system extract S1 and S2 from the mixed signal received by the 16-channel microphone array with the filtered back-projection method we proposed and Figure 9: Simulation of two dual frequencies sources with signal fixed position and noise changed. The higher value delay-and-sum method. Choosing 3 periods of.5 second signal form two result music which is 3 seconds long to calculate the SNR. The comparion between filtered backprojection method and delay-and-sum method is shown in Table 1. 5. Conclusion The simulation result shows the filtered back-projection method may reduce the noise and enhance the signal we wantt. As shown in figures, when the difference of distance from sources to microphone array in a proper range, the signal from the focus source could be extracted clearly. This system could be used in teleconference. Usually, the position of conferees is in front of the screen. According the Eq. 15, microphone array should be arranged around the half room. We consider the realization of the system would be too complex to setup hardware on multi-walls, so only arrange the microphone array under the projection screen. It would be a proper position to receive voice from speaker s height when sitting.

5 Figure 1: Simulation of two single frequency sources with noise fixed position and signal changed. The higher value Figure 11: Simulation of two dual frequencies sources with noise fixed position and signal changed. The higher value Table 1: SNR signal (db) of extraction of music using two methods. The larger value Back- Delay- Time projection -and-sum S 1 S 2 S 1 S 2 t 1 8.34 21.17 1.15 13.26 t 2 18.78 29.6 11.26 13.74 t 3 13.96 9.37 26.19 16.19 t 4 37.18 5.47 1.56 2.13 t 5 1.49 17.48 1.23 12.68 Average 17.13 12.63 References [1] H. F. Silverman, W. R. Patterson, and J. L. Flanagan, The huge microphone array, IEEE Trans. on Concurrency, vol. 6, no. 4, pp. 36 46, Oct.-Dec. 1998. [2] Y. Tamai, S. Kagami, H. Mizoguchi, K. Sakaya, K. Nagashima, and T. Takano, Circular microphone array for meeting system, vol. 2, pp. 11 115. [3] D. Giuliani, M. Matassoni, and M. Omologo, Hands free continuous speech recognition in noisy environment using a four microphone array, in Proc. of IEEE Int l. Conf. on Acoustics, Speech, and Signal Processing 1995 (ICASSP 95), vol. 1, Detroit, MI, May 1995, pp. 86 863. [4] G. T. Herman, Ed., Image Reconstruction from Projections: Implementation and Applications. Berlin: Springer-Verlag, 1979.