Linear and Parametric Microphone Array Processing

Size: px
Start display at page:

Download "Linear and Parametric Microphone Array Processing"

Transcription

1 Part 5 - Joint Linear and Parametric Spatial Processing Emanuël A. P. Habets 1 and Sharon Gannot 2 1 International Audio Laboratories Erlangen, Germany A joint institution of University of Erlangen-Nuremberg and Fraunhofer IIS 2 Faculty of Engineering, Bar-Ilan University, Israel ICASSP 213, Vancouver, Canada

2 Overvie 1 Motivation 2 Informed Spatial Filtering 3 Examples Page 1/31

3 1. Motivation Classical Linear Spatial Filtering: + High amount of noise plus interference reduction + Controllable tradeoff beteen speech distortion and noise reduction + Controllable tradeoff beteen different noise types Not very robust.r.t. estimation errors, position changes, etc. Relatively slo response time Parametric Spatial Filtering: + Fast response time + Relatively robust.r.t. estimation errors, position changes, etc. + Possibility to manipulate parameters (e.g., virtual source displacement) Inherent tradeoff beteen speech distortion and noise reduction Model violations can introduce audible artifacts [Thiergart and Habets, 212] Relatively poor interference reduction due to tradeoff and model violations Page 2/31

4 Overvie 1 Motivation 2 Informed Spatial Filtering 3 Examples Page 3/31

5 2. Informed Spatial Filtering The main idea behind informed filtering is to incorporate relevant information about specific problem into design of filters and estimation of required statistics. Microphone Signals Informed Multichannel Spatial Filter Processed Signals Estimate Second Order Statistics Estimate Parameters (e.g., diffuseness, DOA) Figure: Informed filtering approach. Page 4/31

6 2. Informed Spatial Filtering A selection of parameters that can be used (see Part 4): Signal-to-diffuse ratio (SDR): Γ(k, m, p i) = P dir(k, m, p i) P diff (k, m), here P dir is poer of direct component at position p i and P diff is poer of diffuse component (assuming a ly homogenous sound field). Time and frequency dependent direction-of-arrival estimates. Time and frequency dependent interaural level differences. Time and frequency dependent interaural phase differences.... Page 5/31

7 Overvie 1 Motivation 2 Informed Spatial Filtering 3 Examples Example A: Extracting Coherent Sound Sources Example B: Dereverberation in SH Domain Example C: Directional Filtering Example D: Source Extraction Page 6/31

8 3.1 Example A: Extracting Coherent Sound Sources Signal model: y(k, m) = x(k, m) + v(k, m). Assumption: Desired signals are strongly coherent across array. Aim: Estimate X1(k, m) using a parametric multichannel Wiener filter [Benesty et al., 211]: Φ 1 v (k, m)φ x(k, m) h PMWF(k, m) = λ(k, m) + tr { Φ 1 v (k, m)φ x(k, m) } u1 5 λ Figure: Mapping from input signal-to-diffuse ratio to tradeoff parameter λ [Taseska and Habets, 212]. Page 7/31

9 Proposed Solution [Taseska and Habets, 212] y(k, m) Z(k, m) Parametric Multichannel Wiener Filter Estimate Noise PSD Matrix v(k, m) Estimate Speech Presence Probability P [H 1 y(k, m)] (k, m) Estimate Signal-to- Diffuse Ratio Figure: Block diagram of proposed system. Page 8/31

10 Algorithm Summary High-level description of proposed algorithm [Taseska and Habets, 212]: 1. Compute signal-to-diffuse ratio (SDR) using [Thiergart et al., 212]. 2. Compute a priori speech presence probability (SPP) based on SDR. 3. Compute multichannel a posteriori SPP [Souden et al., 21]. 4. Update noise PSD matrix using a posteriori SPP. 5. Compute tradeoff parameter for parametric multichannel Wiener filter (PMWF) based on SDR: - When SDR is high, e decrease amount of speech distortion. - When SDR is lo, e increase amount of noise reduction. 6. Compute and apply parametric multichannel Wiener filter. Page 9/31

11 Results (1).8 λ = λ = 1 λ = f(γ) λ = f(γ) + SPP.8 Δ PESQ input SNR (db) input SNR (db) Figure: Performance evaluation: PESQ improvement for stationary diffuse noise (left) and diffuse babble speech (right) [Taseska and Habets, 212]. Page 1/31

12 Results (2) λ = λ = 1 λ = f(γ) λ = f(γ) + SPP SNR gain (db) input SNR (db) input SNR (db) Figure: Performance evaluation: segmental SNR improvement for stationary diffuse noise (left) and diffuse babble speech (right) [Taseska and Habets, 212]. Page 11/31

13 Results (3) Frequency [khz] Frequency [khz] Time [s] Time [s] 5 (a) First Microphone Signal (b) MVDR Frequency [khz] Frequency [khz] Time [s] (c) Parametric MWF Time [s] (d) Parametric MWF ith MC-SPP 5 Figure: Examples obtained using M=4 microphone signals corrupted by sensor noise and babble speech (input SNR = 1 db). Audio Examples Page 12/31

14 3.2 Example B: Dereverberation in SH Domain Assumed signal model ith stacked spherical harmonic components: p(k, m) = x(k, m) + d(k, m) + ṽ(k, m) }{{} = γ(k, m) X (k, m) + ũ(k, m) γ(k, m) = x(k, m) X (k, m) = y(ω dir) Y (Ω dir ) = γ dir, here Y is zero-order spherical harmonic and Ω dir is DOA. p d v x Spherical Harmonics up to order 3 Page 13/31

15 Proposed Solution [Braun et al., 213] Desired signal: The direct signal component X(k, m) hich corresponds to sound pressure measured at center of array in absence of spherical microphone array. Assumption: Direct, diffuse and noise components are mutually uncorrelated. Proposed solution: The (rank-1) MWF provides an MMSE estimate of X (k, m). For practical reasons, e split MWF into an MVDR filter folloed by a single-channel Wiener filter: h MWF(k, m) = φ X (k, m) Φ 1 ũ (k, m) γ dir φ X (k, m) γ H dir Φ 1 ũ (k, m) γ dir + 1 φ X = Φ 1 ũ (k, m) γ dir γ H dir Φ 1 ũ (k, m) γ dir φ }{{} X + [ γ H dir Φ 1 ũ (k, m) γ ] 1 dir }{{} h MVDR (k,m) H W (k,m) Page 14/31

16 Parameter-based PSD Matrix Estimation Required information: Diffuse PSD matrix estimation: Direction of arrival (DOA) γ dir Interference PSD matrix: Φũ(k, m) = Φ d(k, m) + Φṽ(k, m) Assume model for diffuse sound component: Φ d(k, m) = φ D (k, m) I (L+1) 2 Calculate diffuse sound PSD using an estimate of diffuseness Ψ: φ D (k, m) = φ P (k, m) φṽ (k, m) Ψ 1 (k, m) STFT SHT STFT -1 Diffuseness Estimation Residual Interf. PSD Diffuse PSD Estimation Page 15/31

17 Results Frequency [khz] Frequency [khz] Time [s] Time [s] 5 (a) Reference X (k, m) (b) Received P (k, m) Frequency [khz] Frequency [khz] Time [s] Time [s] 5 (c) Processed: MVDR (d) Processed: MWF Figure: Examples obtained using simulated signals [Jarrett et al., 212] (source-array distance is 2 m, SNR = 2 db, T 6 =4 ms). Audio examples. Page 16/31

18 3.3 Example C: Directional Filtering Flexible sound acquisition in noisy and reverberant environments ith rapidly changing acoustic scenes is a common problem in modern communication systems. A filter is proposed that provides an arbitrary response for J sources being simultaneously active per time and frequency. The proposed filter provides an optimal tradeoff beteen hite noise gain (WNG) and directivity index. The filter exploits instantaneous information on sound (narroband DOAs, diffuse-to-noise ratio) hich allos a nearly immediate adaption to changes in acoustic scene. Page 17/31

19 Problem Formulation Assuming three components in (1) are mutually uncorrelated, e canthe expressdesired poer signal spectral density is given (PSD) by: matrix of microphone signals as { J } Φ(k, n) Z(k, =E m) x(k, n) = x H (k, n) G(k, ϕ j)x (j) 1 L j=1 = Φl(k, n)+φd(k, n)+φn(k, n), (2) ith Signal model: Based on a multi-ave sound field model, M microphone signals can be expressed as: J y(k, m) = x (j) (k, m) + d(k, m) + v(k, m) }{{}}{{} j=1 }{{} diffuse sound sensor noise J plane aves Aim: Capturing J plane aves (J M) ith desired arbitrary gain hile attenuating sensor noise and reverberation. l=1 (k, m) Φd(k, n) =φd(k, n) Γd(k), (3) Φn(k, n) =φn(k, n) I. (4) Here, I is an identity matrix, φn(k, n) is expected poer of microphone self-noise, hich is identical for all microphones, and φd(k, Linear n) isand Parametric expected poer Microphone of Array diffuse Processing field, hich can vary rapidly Emanuël across Habets time and (FAU) frequency. and Sharon The Gannot ij-th(biu) element of coher- G(k, ϕ) 2 [db] G1 G2 ϕa ϕb DOA ϕ [ ] The desired signal is estimated using an informed LCMV filter: Ẑ(k, m) = h H ilcmv(k, m) y(k, m) 1. Toarbitrarydirectivityfunctions&sourcepositions 3.1. Existing Spatial Filters 3. OPTIMAL SPATIAL FILTERS While PSD φn(k, c International n) can beaudio estimated Laboratories during Erlangen, periods of213 silence, φd(k, n) is commonly assumed unknon and unobservable. Page 18/31 We

20 Proposed Solution (1) The proposed informed LCMV filer is given by: h ilcmv = argmin h h H [Φ d (k, m) + Φ v(k, m)] h s. t. h H (k, m) a(k, ϕ j) = G(k, ϕ j), j {1, 2,..., J} here a(k, ϕ j) denotes steering vector for jth plane ave at time m and frequency k. For assumed signal model, e can alternatively minimize h H [Ψ(k, m) Γ d (k) + I] h, here Ψ(k, m) denotes instantaneous diffuse-to-noise ratio (DNR) and Γ d (k) denotes coherence matrix of diffuse sound field. The filter is updated for each time and frequency given instantaneous parametric information (DOAs, DNR). The filter requires knoledge of DNR, hich can be estimated using an auxiliary filter (see poster session AASP-P8 on Friday or [Thiergart and Habets, 213]). Page 19/31

21 Proposed Solution (2) DI [db] WNG [db] n 5d 4nd nd 6 3 frequency [khz] DNR Estimation DOA Estimation Desired Response 9 45 Filter Weights frequency time [khz] [s] time [s] (b) Mean (a) DOA WNG ϕ1(k, n) [ ] (b) G(k, ϕ1) 2 [db] 3. DI and WNG of filters in Sec. 3. For d, Figure: Left: DOA ϕ minimum WNG 5 1 (k, m) as a function of time 4. EstimatedDOAϕ1(k, and frequency. n) Right: and resulting Desired gains G(k, ϕ1) as set to 12 db to make filter robust response G(k, ϕ 5 against microphone 4 self-noise. 1 ) 2 in db for DOA ϕ 1 (k, m) as a function of time and frequency. fit. The best performance in terms of SSNR is obtained using n. 3 1 In terms of PESQ, nd and d outperform n. Usinginstanta- neous directional c constraints International (asaudio in this Laboratories section) instead Erlangen, of 213 time- for alinear silent part and 2 Parametric of signal Microphone and duringarray speech Processing activity [both signal 15 partsemanuël marked in 2(b)]. During silence, proposed filter 1 Habets (FAU) and Sharon Gannot (BIU) invariant constrains (as in Sec. 5.1, values in brackets) Page mainly 2/31 re- frequency [khz] n d nd nd 1 frequency [khz] (a) Mean DI frequency [khz] frequency [khz] time [s] (a) DOA ϕ1(k, n) [ ]

22 Informed InformedLCMV LCMVFilter Filter Results (1) DIDI[dB] [db] DIDI[dB] [db].(9). t. (9). (12) (12) pressed essed as as nn dn dn dnd dnd nd nd nd nd ndnd sources active sources active sources silent sources silent time time [s] [s] True Ψ(k, [db] (a)(a) True Ψ(k, n) n) [db] nes. The solu. The solu frequency [khz] frequency [khz] (14) (14) (15) (15) WNG WNG[dB] [db] WNG WNG[dB] [db] t.9),(9), (13) (13) frequency [khz] frequency [khz] frequency [khz] frequency [khz] Mean (a)(a) Mean DIDI Mean (a)(a) Mean DIDI nn dn dn nd nd d d nd nd nd nd ndnd sources silent sources silent sources active sources active frequency [khz] frequency [khz] frequency frequency[khz] [khz] frequency frequency[khz] [khz] 7 7 frequency [khz] frequency [khz] time time [s] [s] he propagation propagation Mean WNG (b)(b) Mean WNG Estimated Ψ(k, [db] (b)(b) Estimated Ψ(k, n) n) [db] Mean WNG (b)(b) Mean WNG ns given areare given byby WNG filters Sec., 3. 3.DIDI andand WNG of of filters in in Sec ForFor d, d nf of Ψ(k, Ψ(k, n)n) is is 3.DIDI and WNG of filters Sec., True estimated DNR Ψ(k, The to marked areas 3. and WNG of filters in in Sec ForFor d, robust 2. 2.True andand estimated DNR Ψ(k, n).n).the to marked areas d minimum WNG as 12 make filter minimum WNG as setset to to 12 dbdb to to make filter robust Figure: Top: Directivity index (DI) in db. Figure: Top: True DNR in db. Bottom: indicate respectively a silent and active part signal. indicate respectively a silent and active part of of signal. minimum WNG as set to 12 db to make filter robust minimum WNG as set to 12 db to make filter robust against microphone self-noise. against microphone self-noise. against microphone self-noise. against microphone self-noise. Bottom: White noise gain (WNG) in db. Estimated DNR in db. frequency frequency[khz] [khz] frequency frequency[khz] [khz] frequency [khz] frequency [khz] lem m in in (8)(8) areare er diffuse andand diffuse Es Estim Es Estim Table 1 Table 1 su Table 1 Table 1 su International Erlangen International AudioLabs Erlangen terms in in terms of of sigs The required expected poer microphone self-noise n) minimizes The required expected poer of of microphone self-noise φn φ (k, n)audiolabs n (k, noise poer, The in terms nactivity terms of of sigs speech activity reverberant environment. The estimated in destimated speech duedue to to reverberant environment. ratio (SRR) ratio (SRR), a example estimated during silence assuming that poer cancan forfor example be be estimated during silence assuming that poer 6 activity 6speech activity due to reverberant environment. The estimated speech due to reverberant environment. The estimated ratio (SRR) (SRR), DNR 2(b) possesses a limited temporal resolution due to DNR in in 2(b) possesses a limited temporal resolution due to ratio ed approach approach in in minimizes diffuse poer, is SRR (sourc SRR (source a nd duedue is constant over time.note Note that proposed DNR estimator does is constant over time. that proposed DNR estimator does DNR 2(b) possesses a limited temporal resolution DNR in in 2(b) possesses a limited temporal resolution to to SRR (sourc (source an obtained incorporated temporal averaging process. Neverless, Ψ(k, be be obtained incorporated temporal averaging process. Neverless, Ψ(k, n)n) SRR ( ) and th ( ) and necessarily provide loest estimation variance practice nd notnot necessarily provide loest estimation variance in in practice nd incorporated temporal averaging process. Neverless, Ψ(k, n) incorporated temporal averaging process. Neverless, Ψ(k, n) proposed ilcmv filter that minimizes (( ) and th ssprit ESPRIT [24] [24] estimates sufficiently accurate shon folloing results. estimates areare sufficiently accurate as as shon byby folloing results. nd )ndand formance. HoH chosen optimization criteria (16), provides unbiased duedue to to chosen optimization criteria (16), butbut provides unbiased estimates sufficiently accurate shon folloing results. formance. estimates areare sufficiently accurate as as shon byby folloing results. formance.hoh estimation stimation of of Figure 3(a) depicts mean and both formance. Figure 3(a) depicts mean DIDI forfor n[shon d(hich areare both diffuse plus noise poer hen n and d (hich than SS SSNR results. results. Figure 3(a) depicts mean DI (hich are both than Figure 3(a) depicts for mean DIproposed forfor n and dfilter both n and d (hich than SS than SSNR signal-independent), for filter are (hich signal-independent), andand proposed nd (hich nd The bes fit.fit.the best p sources are active (red solid line) and signal-independent), and for proposed filter (hich signal-independent), and for proposed filter (hich nd nd onal fit.the The besp l fil-filfit.in best is signal-dependent). proposed filter, sho is signal-dependent). ForFor proposed filter, ee sho DIDI In terms of terms of P is signal-dependent). proposed filter, sho is signal-dependent). ForFor proposed filter, ee sho DIDI In In EXPERIMENTAL RESULTS 5. 5.EXPERIMENTAL RESULTS ffuse sound use sound is is terms of terms of P silent (red dashed line)]. a silent part signal during speech activity [both signal forfor a silent part of of signal andand during speech activity [both signal neousdirectio direc neous for a silent part of signal and during speech activity [both signal for a silent part of signal and during speech activity [both signal by maximizing maximizing neousdirectio direc neous parts marked 2(b)]. During silence, proposed filter invariant parts marked in in 2(b)]. During silence, proposed filter invariant co cons assume L= 2 plane aves model ULA LetLet us us assume L= 2 plane aves in in model in in (1)(1) andand an an ULA parts marked in 2(b)]. During silence, proposed filter invariant parts marked in During silence, filter invariant co cons (dashed line ) 2(b)]. provides same lo DI as. During speech (dashed line same lo DI asproposed n. nduring speech nd )ndprovides duced duced acha ith M 4 microphones ith an inter-microphone spacing 3 cm. ith MParametric = 4=microphones ith anarray inter-microphone of of 3 cm. (dashed line )ndprovides ) provides same lo as. During speech213 (dashed line nd same lo DIisDI as. nduring speech c), International Linear and Microphone Processing spacing Audio Laboratories Erlangen, n duced duced acha activity (solid line ), obtained DI as high as for robust activity (solid line obtained DI is as high as for robust 3 3 nd nd varying sou varying source reverberant shoebox room (7., RT 38 ms) AA reverberant shoebox room ( mm, RT 38 ms) 6 6 activity (solid line ),nd ), obtained DI is as high as for robust (16) activity (solid line obtained DI is as high as for robust (16) nd varying sou varying source Emanue l Habets (FAU) and Sharon Gannot (BIU) Page 21/31 beamformer ). Figure 3(b) shos corresponding WNGs. quired SDSD beamformer (( 3(b) shos corresponding WNGs. d ).dfigure quired comp comple as simulated using source-image method [26, ith to as simulated using source-image method [26, 27]27] ith to beamformer ). Figure 3(b) shos corresponding WNGs. SDSD beamformer (( ). Figure 3(b) shos corresponding WNGs.

23 Results (2) The proposed filter provides a high DI hen sound field is diffuse and a high WNG hen sensor noise is dominant. Interfering sound can be strongly attenuated if desired. The proposed DNR estimator provides a sufficiently high accuracy and temporal resolution to allo signal enhancement under adverse conditions even in changing acoustic scenes. SegSIR [db] SegSRR [db] SegSNR [db] PESQ 11 (11) 7 ( 7) 26 (26) 1.5 (1.5) n 21 (32) 2 ( 3) 33 (31) 2. (1.7) d 26 (35) ( 1) 22 (24) 2.1 (2.) nd 25 (35) 1 ( 1) 28 (26) 2.1 (2.) Table: Performance of all filters [ unprocessed, first sub-column using true DOAs (of sources), second sub-column using estimated DOAs (of plane aves)]. Audio Examples Page 22/31

24 3.4 Example D: Source Extraction Scenario Multiple talkers Additive background noise Distributed sensor arrays Applications Spatial filter Estimate of desired source at microphone Teleconferencing systems Automatic speech recognition Spatial sound reproduction Signal model: y(k, m) = x (d) (k, m) + i d x (i) (k, m) + v(k, m). (d) Aim: Obtain an MMSE estimate of X 1 (k, m). Page 23/31

25 Proposed Solution [Taseska and Habets, 213] Hyposes: H v : y(k, m) = v(k, m) speech absent H x : y(k, m) = x(k, m) + v(k, m) speech present J Hx j : y(k, m) = x (j) (k, m) + x (i) (k, m) + v(k, m) j = 1, 2,..., J i j } {{ } Recursive estimation of PSD matrices: ( Φ (j) x (m) = p[hx j y] α Φ(j) x x (m 1) + (1 α x) yy H) ( + 1 p[hx j y] ) Φ(j) x (m 1) Signal-to-diffuse ratio (Γ) and position (Θ) -based posterior probabilities: p[h j x y] = p[h j x y, H x] p[h x y] p[h j x Θ, H x] p[h x Γ, y] Page 24/31

26 Parameter-based PSD Matrix Estimation Γ p[h x Γ, y] Φv Ω Θ p[h j x Θ, H x] ˆp[Hj x y] Φ (j) x The distribution p[ Θ Hx] is modelled as a Gaussian mixture (GM). GM parameters estimated by Expectation-Maximization algorithm. Page 25/31

27 Results (1) Setup: Three reverberant sources ith approximately equal poer, diffuse babble speech (SNR=22 db), and uncorrelated sensor noise (SNR =5 db). The reverberation time as T6 = 25 ms. To uniform circular arrays ere used ith three omnidirectional microphones, a diameter 2.5 cm and an inter-array spacing of 1.5 m. (a) Training during single-talk (b) Training during triple-talk Figure: Output of EM algorithm (3 iterations) and 4.5 s of noisy speech data. The actual source positions are denoted by hite squares. The array location is marked by a plus symbol. The interior of each ellipse contains 85% probability mass of respective Gaussian. Page 26/31

28 Results (2) mixture mixture reference source signals reference source signals (1) (2) (3) extracted source signals extracted source signals (1) (2) (3) time (s) time (s) Figure: Left: constant triple-talk scenario. Right: mainly single-talk scenario. (S,M) Audio files available at Page 27/31

29 More Information These and or examples are presented at ICASSP 213 on: Friday 1:3-12:3 in Poster Session AASP-P8: An Informed Spatial Filter in Spherical Harmonic Domain for Joint Noise Reduction and Dereverberation (Braun, Jarret, Fischer and Habets) Friday 1:3-12:3 in Poster Session AASP-P8: An Informed LCMV Filter Based on Multiple Instantaneous Direction-Of-Arrival Estimates (Thiergart and Habets) Friday 1:3-12:3 in Poster Session AASP-P8 MMSE-based Source Extraction using Position-based Posterior Probabilities (Taseska and Habets) Friday 1:3-12:3 in Poster Session AASP-P8 Spherical Harmonic Domain Noise Reduction Using an MVDR Beamformer and DOA-based Second-order Statistics Estimation (Jarrett, Habets and Naylor) Page 28/31

30 Special thanks to Sebastian Braun, Maja Taseska, Oliver Thiergart and Daniel Jarrett for ir contributions. Page 29/31

31 References I Benesty, J., Chen, J., and Habets, E. A. P. (211). Speech Enhancement in STFT Domain. SpringerBriefs in Electrical and Computer Engineering. Springer-Verlag. Braun, S., Jarrett, D. P., Fischer, J., and Habets, E. A. P. (213). An informed filter for dereverberation in spherical harmonic domain. In Proc. IEEE Intl. Conf. on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, Canada. Jarrett, D. P., Habets, E. A. P., Thomas, M. R. P., and Naylor, P. A. (212). Rigid sphere room impulse response simulation: algorithm and applications. J. Acoust. Soc. Am., 132(3): Souden, M., Chen, J., Benesty, J., and Affes, S. (21). Gaussian model-based multichannel speech presence probability. IEEE Trans. Audio, Speech, Lang. Process., 18(5): Taseska, M. and Habets, E. (213). MMSE-based source extraction using position-based posterior probabilities. In Proc. IEEE Intl. Conf. on Acoustics, Speech and Signal Processing (ICASSP). Page 3/31

32 References II Taseska, M. and Habets, E. A. P. (212). MMSE-based blind source extraction in diffuse noise fields using a complex coherence-based a priori SAP estimator. In Proc. Intl. Workshop Acoust. Signal Enhancement (IWAENC). Thiergart, O., Del Galdo, G., and Habets, E. A. P. (212). On coherence in mixed sound fields and its application to signal-to-diffuse ratio estimation. J. Acoust. Soc. Am., 132(4): Thiergart, O. and Habets, E. (213). Informed optimum filtering using multiple instantaneous direction-of-arrival estimates. In Proc. IEEE Intl. Conf. on Acoustics, Speech and Signal Processing (ICASSP). Thiergart, O. and Habets, E. A. P. (212). Sound field model violations in parametric sound processing. In Proc. Intl. Workshop Acoust. Signal Enhancement (IWAENC). Page 31/31

Recent Advances in Acoustic Signal Extraction and Dereverberation

Recent Advances in Acoustic Signal Extraction and Dereverberation Recent Advances in Acoustic Signal Extraction and Dereverberation Emanuël Habets Erlangen Colloquium 2016 Scenario Spatial Filtering Estimated Desired Signal Undesired sound components: Sensor noise Competing

More information

Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor. Presented by Amir Kiperwas

Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor. Presented by Amir Kiperwas Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor Presented by Amir Kiperwas 1 M-element microphone array One desired source One undesired source Ambient noise field Signals: Broadband Mutually

More information

Informed Spatial Filtering for Sound Extraction Using Distributed Microphone Arrays

Informed Spatial Filtering for Sound Extraction Using Distributed Microphone Arrays IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 7, JULY 2014 1195 Informed Spatial Filtering for Sound Extraction Using Distributed Microphone Arrays Maja Taseska, Student

More information

arxiv: v1 [cs.sd] 4 Dec 2018

arxiv: v1 [cs.sd] 4 Dec 2018 LOCALIZATION AND TRACKING OF AN ACOUSTIC SOURCE USING A DIAGONAL UNLOADING BEAMFORMING AND A KALMAN FILTER Daniele Salvati, Carlo Drioli, Gian Luca Foresti Department of Mathematics, Computer Science and

More information

EUSIPCO

EUSIPCO EUSIPCO 97 AN INFORMED MMSE FILTER BASED ON MULTIPLE INSTANTANEOUS DIRECTION-OF-ARRIVAL ESTIMATES Oliver Thiergart, Maja Taseska, and Emanuël A. P. Habets International Audio Laboratories Erlangen Am Wolfsmantel,

More information

Blind Beamforming for Cyclostationary Signals

Blind Beamforming for Cyclostationary Signals Course Page 1 of 12 Submission date: 13 th December, Blind Beamforming for Cyclostationary Signals Preeti Nagvanshi Aditya Jagannatham UCSD ECE Department 9500 Gilman Drive, La Jolla, CA 92093 Course Project

More information

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 5, MAY

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 5, MAY IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 5, MAY 2013 945 A Two-Stage Beamforming Approach for Noise Reduction Dereverberation Emanuël A. P. Habets, Senior Member, IEEE,

More information

Speech Enhancement Using Microphone Arrays

Speech Enhancement Using Microphone Arrays Friedrich-Alexander-Universität Erlangen-Nürnberg Lab Course Speech Enhancement Using Microphone Arrays International Audio Laboratories Erlangen Prof. Dr. ir. Emanuël A. P. Habets Friedrich-Alexander

More information

A BROADBAND BEAMFORMER USING CONTROLLABLE CONSTRAINTS AND MINIMUM VARIANCE

A BROADBAND BEAMFORMER USING CONTROLLABLE CONSTRAINTS AND MINIMUM VARIANCE A BROADBAND BEAMFORMER USING CONTROLLABLE CONSTRAINTS AND MINIMUM VARIANCE Sam Karimian-Azari, Jacob Benesty,, Jesper Rindom Jensen, and Mads Græsbøll Christensen Audio Analysis Lab, AD:MT, Aalborg University,

More information

Microphone Array Design and Beamforming

Microphone Array Design and Beamforming Microphone Array Design and Beamforming Heinrich Löllmann Multimedia Communications and Signal Processing heinrich.loellmann@fau.de with contributions from Vladi Tourbabin and Hendrik Barfuss EUSIPCO Tutorial

More information

ROBUST SUPERDIRECTIVE BEAMFORMER WITH OPTIMAL REGULARIZATION

ROBUST SUPERDIRECTIVE BEAMFORMER WITH OPTIMAL REGULARIZATION ROBUST SUPERDIRECTIVE BEAMFORMER WITH OPTIMAL REGULARIZATION Aviva Atkins, Yuval Ben-Hur, Israel Cohen Department of Electrical Engineering Technion - Israel Institute of Technology Technion City, Haifa

More information

A MULTI-CHANNEL POSTFILTER BASED ON THE DIFFUSE NOISE SOUND FIELD. Lukas Pfeifenberger 1 and Franz Pernkopf 1

A MULTI-CHANNEL POSTFILTER BASED ON THE DIFFUSE NOISE SOUND FIELD. Lukas Pfeifenberger 1 and Franz Pernkopf 1 A MULTI-CHANNEL POSTFILTER BASED ON THE DIFFUSE NOISE SOUND FIELD Lukas Pfeifenberger 1 and Franz Pernkopf 1 1 Signal Processing and Speech Communication Laboratory Graz University of Technology, Graz,

More information

A Three-Microphone Adaptive Noise Canceller for Minimizing Reverberation and Signal Distortion

A Three-Microphone Adaptive Noise Canceller for Minimizing Reverberation and Signal Distortion American Journal of Applied Sciences 5 (4): 30-37, 008 ISSN 1546-939 008 Science Publications A Three-Microphone Adaptive Noise Canceller for Minimizing Reverberation and Signal Distortion Zayed M. Ramadan

More information

Joint recognition and direction-of-arrival estimation of simultaneous meetingroom acoustic events

Joint recognition and direction-of-arrival estimation of simultaneous meetingroom acoustic events INTERSPEECH 2013 Joint recognition and direction-of-arrival estimation of simultaneous meetingroom acoustic events Rupayan Chakraborty and Climent Nadeu TALP Research Centre, Department of Signal Theory

More information

Dual-Microphone Speech Dereverberation in a Noisy Environment

Dual-Microphone Speech Dereverberation in a Noisy Environment Dual-Microphone Speech Dereverberation in a Noisy Environment Emanuël A. P. Habets Dept. of Electrical Engineering Technische Universiteit Eindhoven Eindhoven, The Netherlands Email: e.a.p.habets@tue.nl

More information

Microphone Array Power Ratio for Speech Quality Assessment in Noisy Reverberant Environments 1

Microphone Array Power Ratio for Speech Quality Assessment in Noisy Reverberant Environments 1 for Speech Quality Assessment in Noisy Reverberant Environments 1 Prof. Israel Cohen Department of Electrical Engineering Technion - Israel Institute of Technology Technion City, Haifa 3200003, Israel

More information

Enhancement of Speech Signal Based on Improved Minima Controlled Recursive Averaging and Independent Component Analysis

Enhancement of Speech Signal Based on Improved Minima Controlled Recursive Averaging and Independent Component Analysis Enhancement of Speech Signal Based on Improved Minima Controlled Recursive Averaging and Independent Component Analysis Mohini Avatade & S.L. Sahare Electronics & Telecommunication Department, Cummins

More information

Dual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation

Dual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation Dual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation Gal Reuven Under supervision of Sharon Gannot 1 and Israel Cohen 2 1 School of Engineering, Bar-Ilan University,

More information

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B. www.ijecs.in International Journal Of Engineering And Computer Science ISSN:2319-7242 Volume 4 Issue 4 April 2015, Page No. 11143-11147 Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya

More information

NOISE POWER SPECTRAL DENSITY MATRIX ESTIMATION BASED ON MODIFIED IMCRA. Qipeng Gong, Benoit Champagne and Peter Kabal

NOISE POWER SPECTRAL DENSITY MATRIX ESTIMATION BASED ON MODIFIED IMCRA. Qipeng Gong, Benoit Champagne and Peter Kabal NOISE POWER SPECTRAL DENSITY MATRIX ESTIMATION BASED ON MODIFIED IMCRA Qipeng Gong, Benoit Champagne and Peter Kabal Department of Electrical & Computer Engineering, McGill University 3480 University St.,

More information

MULTICHANNEL AUDIO DATABASE IN VARIOUS ACOUSTIC ENVIRONMENTS

MULTICHANNEL AUDIO DATABASE IN VARIOUS ACOUSTIC ENVIRONMENTS MULTICHANNEL AUDIO DATABASE IN VARIOUS ACOUSTIC ENVIRONMENTS Elior Hadad 1, Florian Heese, Peter Vary, and Sharon Gannot 1 1 Faculty of Engineering, Bar-Ilan University, Ramat-Gan, Israel Institute of

More information

OPTIMUM POST-FILTER ESTIMATION FOR NOISE REDUCTION IN MULTICHANNEL SPEECH PROCESSING

OPTIMUM POST-FILTER ESTIMATION FOR NOISE REDUCTION IN MULTICHANNEL SPEECH PROCESSING 14th European Signal Processing Conference (EUSIPCO 6), Florence, Italy, September 4-8, 6, copyright by EURASIP OPTIMUM POST-FILTER ESTIMATION FOR NOISE REDUCTION IN MULTICHANNEL SPEECH PROCESSING Stamatis

More information

Flexible and efficient spatial sound acquisition and subsequent. Parametric Spatial Sound Processing

Flexible and efficient spatial sound acquisition and subsequent. Parametric Spatial Sound Processing [ Konrad Kowalczyk, Oliver Thiergart, Maja Taseska, Giovanni Del Galdo, Ville Pulkki, and Emanuël A.P. Habets ] Parametric Spatial Sound Processing ear photo istockphoto.com/xrender assisted listening

More information

260 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 2, FEBRUARY /$ IEEE

260 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 2, FEBRUARY /$ IEEE 260 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 2, FEBRUARY 2010 On Optimal Frequency-Domain Multichannel Linear Filtering for Noise Reduction Mehrez Souden, Student Member,

More information

The Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals

The Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals The Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals Maria G. Jafari and Mark D. Plumbley Centre for Digital Music, Queen Mary University of London, UK maria.jafari@elec.qmul.ac.uk,

More information

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 24, NO. 7, JULY

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 24, NO. 7, JULY IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 24, NO. 7, JULY 2016 1291 Spotforming: Spatial Filtering With Distributed Arrays for Position-Selective Sound Acquisition Maja Taseska,

More information

Performance Evaluation of Nonlinear Speech Enhancement Based on Virtual Increase of Channels in Reverberant Environments

Performance Evaluation of Nonlinear Speech Enhancement Based on Virtual Increase of Channels in Reverberant Environments Performance Evaluation of Nonlinear Speech Enhancement Based on Virtual Increase of Channels in Reverberant Environments Kouei Yamaoka, Shoji Makino, Nobutaka Ono, and Takeshi Yamada University of Tsukuba,

More information

HUMAN speech is frequently encountered in several

HUMAN speech is frequently encountered in several 1948 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 20, NO. 7, SEPTEMBER 2012 Enhancement of Single-Channel Periodic Signals in the Time-Domain Jesper Rindom Jensen, Student Member,

More information

Performance Evaluation of Capon and Caponlike Algorithm for Direction of Arrival Estimation

Performance Evaluation of Capon and Caponlike Algorithm for Direction of Arrival Estimation Performance Evaluation of Capon and Caponlike Algorithm for Direction of Arrival Estimation M H Bhede SCOE, Pune, D G Ganage SCOE, Pune, Maharashtra, India S A Wagh SITS, Narhe, Pune, India Abstract: Wireless

More information

IN REVERBERANT and noisy environments, multi-channel

IN REVERBERANT and noisy environments, multi-channel 684 IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 11, NO. 6, NOVEMBER 2003 Analysis of Two-Channel Generalized Sidelobe Canceller (GSC) With Post-Filtering Israel Cohen, Senior Member, IEEE Abstract

More information

/$ IEEE

/$ IEEE IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 6, AUGUST 2009 1071 Multichannel Eigenspace Beamforming in a Reverberant Noisy Environment With Multiple Interfering Speech Signals

More information

Multiple Sound Sources Localization Using Energetic Analysis Method

Multiple Sound Sources Localization Using Energetic Analysis Method VOL.3, NO.4, DECEMBER 1 Multiple Sound Sources Localization Using Energetic Analysis Method Hasan Khaddour, Jiří Schimmel Department of Telecommunications FEEC, Brno University of Technology Purkyňova

More information

Adaptive Beamforming Applied for Signals Estimated with MUSIC Algorithm

Adaptive Beamforming Applied for Signals Estimated with MUSIC Algorithm Buletinul Ştiinţific al Universităţii "Politehnica" din Timişoara Seria ELECTRONICĂ şi TELECOMUNICAŢII TRANSACTIONS on ELECTRONICS and COMMUNICATIONS Tom 57(71), Fascicola 2, 2012 Adaptive Beamforming

More information

DISTANT or hands-free audio acquisition is required in

DISTANT or hands-free audio acquisition is required in 158 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 1, JANUARY 2010 New Insights Into the MVDR Beamformer in Room Acoustics E. A. P. Habets, Member, IEEE, J. Benesty, Senior Member,

More information

AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION

AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION Gerhard Doblinger Institute of Communications and Radio-Frequency Engineering Vienna University of Technology Gusshausstr. 5/39,

More information

AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION

AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION 1th European Signal Processing Conference (EUSIPCO ), Florence, Italy, September -,, copyright by EURASIP AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION Gerhard Doblinger Institute

More information

SPECTRAL COMBINING FOR MICROPHONE DIVERSITY SYSTEMS

SPECTRAL COMBINING FOR MICROPHONE DIVERSITY SYSTEMS 17th European Signal Processing Conference (EUSIPCO 29) Glasgow, Scotland, August 24-28, 29 SPECTRAL COMBINING FOR MICROPHONE DIVERSITY SYSTEMS Jürgen Freudenberger, Sebastian Stenzel, Benjamin Venditti

More information

BEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR

BEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR BeBeC-2016-S9 BEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR Clemens Nau Daimler AG Béla-Barényi-Straße 1, 71063 Sindelfingen, Germany ABSTRACT Physically the conventional beamforming method

More information

Antennas and Propagation. Chapter 5c: Array Signal Processing and Parametric Estimation Techniques

Antennas and Propagation. Chapter 5c: Array Signal Processing and Parametric Estimation Techniques Antennas and Propagation : Array Signal Processing and Parametric Estimation Techniques Introduction Time-domain Signal Processing Fourier spectral analysis Identify important frequency-content of signal

More information

Joint dereverberation and residual echo suppression of speech signals in noisy environments Habets, E.A.P.; Gannot, S.; Cohen, I.; Sommen, P.C.W.

Joint dereverberation and residual echo suppression of speech signals in noisy environments Habets, E.A.P.; Gannot, S.; Cohen, I.; Sommen, P.C.W. Joint dereverberation and residual echo suppression of speech signals in noisy environments Habets, E.A.P.; Gannot, S.; Cohen, I.; Sommen, P.C.W. Published in: IEEE Transactions on Audio, Speech, and Language

More information

Single channel noise reduction

Single channel noise reduction Single channel noise reduction Basics and processing used for ETSI STF 94 ETSI Workshop on Speech and Noise in Wideband Communication Claude Marro France Telecom ETSI 007. All rights reserved Outline Scope

More information

WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY

WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY INTER-NOISE 216 WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY Shumpei SAKAI 1 ; Tetsuro MURAKAMI 2 ; Naoto SAKATA 3 ; Hirohumi NAKAJIMA 4 ; Kazuhiro NAKADAI

More information

MULTICHANNEL systems are often used for

MULTICHANNEL systems are often used for IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 52, NO. 5, MAY 2004 1149 Multichannel Post-Filtering in Nonstationary Noise Environments Israel Cohen, Senior Member, IEEE Abstract In this paper, we present

More information

Adaptive Beamforming. Chapter Signal Steering Vectors

Adaptive Beamforming. Chapter Signal Steering Vectors Chapter 13 Adaptive Beamforming We have already considered deterministic beamformers for such applications as pencil beam arrays and arrays with controlled sidelobes. Beamformers can also be developed

More information

ONE of the most common and robust beamforming algorithms

ONE of the most common and robust beamforming algorithms TECHNICAL NOTE 1 Beamforming algorithms - beamformers Jørgen Grythe, Norsonic AS, Oslo, Norway Abstract Beamforming is the name given to a wide variety of array processing algorithms that focus or steer

More information

546 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 4, MAY /$ IEEE

546 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 4, MAY /$ IEEE 546 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL 17, NO 4, MAY 2009 Relative Transfer Function Identification Using Convolutive Transfer Function Approximation Ronen Talmon, Israel

More information

Speech Enhancement for Nonstationary Noise Environments

Speech Enhancement for Nonstationary Noise Environments Signal & Image Processing : An International Journal (SIPIJ) Vol., No.4, December Speech Enhancement for Nonstationary Noise Environments Sandhya Hawaldar and Manasi Dixit Department of Electronics, KIT

More information

Design of Robust Differential Microphone Arrays

Design of Robust Differential Microphone Arrays IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 10, OCTOBER 2014 1455 Design of Robust Differential Microphone Arrays Liheng Zhao, Jacob Benesty, Jingdong Chen, Senior Member,

More information

Effective post-processing for single-channel frequency-domain speech enhancement Weifeng Li a

Effective post-processing for single-channel frequency-domain speech enhancement Weifeng Li a R E S E A R C H R E P O R T I D I A P Effective post-processing for single-channel frequency-domain speech enhancement Weifeng Li a IDIAP RR 7-7 January 8 submitted for publication a IDIAP Research Institute,

More information

Direction-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method

Direction-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method Direction-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method Udo Klein, Member, IEEE, and TrInh Qu6c VO School of Electrical Engineering, International University,

More information

LETTER Pre-Filtering Algorithm for Dual-Microphone Generalized Sidelobe Canceller Using General Transfer Function

LETTER Pre-Filtering Algorithm for Dual-Microphone Generalized Sidelobe Canceller Using General Transfer Function IEICE TRANS. INF. & SYST., VOL.E97 D, NO.9 SEPTEMBER 2014 2533 LETTER Pre-Filtering Algorithm for Dual-Microphone Generalized Sidelobe Canceller Using General Transfer Function Jinsoo PARK, Wooil KIM,

More information

COMPARISON OF TWO BINAURAL BEAMFORMING APPROACHES FOR HEARING AIDS

COMPARISON OF TWO BINAURAL BEAMFORMING APPROACHES FOR HEARING AIDS COMPARISON OF TWO BINAURAL BEAMFORMING APPROACHES FOR HEARING AIDS Elior Hadad, Daniel Marquardt, Wenqiang Pu 3, Sharon Gannot, Simon Doclo, Zhi-Quan Luo, Ivo Merks 5 and Tao Zhang 5 Faculty of Engineering,

More information

A BINAURAL HEARING AID SPEECH ENHANCEMENT METHOD MAINTAINING SPATIAL AWARENESS FOR THE USER

A BINAURAL HEARING AID SPEECH ENHANCEMENT METHOD MAINTAINING SPATIAL AWARENESS FOR THE USER A BINAURAL EARING AID SPEEC ENANCEMENT METOD MAINTAINING SPATIAL AWARENESS FOR TE USER Joachim Thiemann, Menno Müller and Steven van de Par Carl-von-Ossietzky University Oldenburg, Cluster of Excellence

More information

Chapter 4 SPEECH ENHANCEMENT

Chapter 4 SPEECH ENHANCEMENT 44 Chapter 4 SPEECH ENHANCEMENT 4.1 INTRODUCTION: Enhancement is defined as improvement in the value or Quality of something. Speech enhancement is defined as the improvement in intelligibility and/or

More information

JOINT DOA AND FUNDAMENTAL FREQUENCY ESTIMATION METHODS BASED ON 2-D FILTERING

JOINT DOA AND FUNDAMENTAL FREQUENCY ESTIMATION METHODS BASED ON 2-D FILTERING 18th European Signal Processing Conference (EUSIPCO-20) Aalborg, Denmark, August 23-27, 20 JOINT DOA AND FUNDAMENTA FREQUENCY ESTIMATION METHODS BASED ON 2-D FITERING Jesper Rindom Jensen, Mads Græsbøll

More information

Calibration of Microphone Arrays for Improved Speech Recognition

Calibration of Microphone Arrays for Improved Speech Recognition MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Calibration of Microphone Arrays for Improved Speech Recognition Michael L. Seltzer, Bhiksha Raj TR-2001-43 December 2001 Abstract We present

More information

IN DISTANT speech communication scenarios, where the

IN DISTANT speech communication scenarios, where the IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 26, NO. 6, JUNE 2018 1119 Linear Prediction-Based Online Dereverberation and Noise Reduction Using Alternating Kalman Filters Sebastian

More information

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter 1 Gupteswar Sahu, 2 D. Arun Kumar, 3 M. Bala Krishna and 4 Jami Venkata Suman Assistant Professor, Department of ECE,

More information

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming Gerhard Schmidt Christian-Albrechts-Universität zu Kiel Faculty of Engineering Electrical Engineering and Information Engineering

More information

Blind Dereverberation of Single-Channel Speech Signals Using an ICA-Based Generative Model

Blind Dereverberation of Single-Channel Speech Signals Using an ICA-Based Generative Model Blind Dereverberation of Single-Channel Speech Signals Using an ICA-Based Generative Model Jong-Hwan Lee 1, Sang-Hoon Oh 2, and Soo-Young Lee 3 1 Brain Science Research Center and Department of Electrial

More information

NOISE ESTIMATION IN A SINGLE CHANNEL

NOISE ESTIMATION IN A SINGLE CHANNEL SPEECH ENHANCEMENT FOR CROSS-TALK INTERFERENCE by Levent M. Arslan and John H.L. Hansen Robust Speech Processing Laboratory Department of Electrical Engineering Box 99 Duke University Durham, North Carolina

More information

arxiv: v1 [cs.sd] 17 Dec 2018

arxiv: v1 [cs.sd] 17 Dec 2018 CIRCULAR STATISTICS-BASED LOW COMPLEXITY DOA ESTIMATION FOR HEARING AID APPLICATION L. D. Mosgaard, D. Pelegrin-Garcia, T. B. Elmedyb, M. J. Pihl, P. Mowlaee Widex A/S, Nymøllevej 6, DK-3540 Lynge, Denmark

More information

SPEECH signals are inherently sparse in the time and frequency

SPEECH signals are inherently sparse in the time and frequency IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 7, SEPTEMBER 2011 2159 An Integrated Solution for Online Multichannel Noise Tracking Reduction Mehrez Souden, Member, IEEE, Jingdong

More information

Recent advances in noise reduction and dereverberation algorithms for binaural hearing aids

Recent advances in noise reduction and dereverberation algorithms for binaural hearing aids Recent advances in noise reduction and dereverberation algorithms for binaural hearing aids Prof. Dr. Simon Doclo University of Oldenburg, Dept. of Medical Physics and Acoustics and Cluster of Excellence

More information

Dual-Microphone Speech Dereverberation using a Reference Signal Habets, E.A.P.; Gannot, S.

Dual-Microphone Speech Dereverberation using a Reference Signal Habets, E.A.P.; Gannot, S. DualMicrophone Speech Dereverberation using a Reference Signal Habets, E.A.P.; Gannot, S. Published in: Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP

More information

Broadband Microphone Arrays for Speech Acquisition

Broadband Microphone Arrays for Speech Acquisition Broadband Microphone Arrays for Speech Acquisition Darren B. Ward Acoustics and Speech Research Dept. Bell Labs, Lucent Technologies Murray Hill, NJ 07974, USA Robert C. Williamson Dept. of Engineering,

More information

PATH UNCERTAINTY ROBUST BEAMFORMING. Richard Stanton and Mike Brookes. Imperial College London {rs408,

PATH UNCERTAINTY ROBUST BEAMFORMING. Richard Stanton and Mike Brookes. Imperial College London {rs408, PATH UNCERTAINTY ROBUST BEAMFORMING Richard Stanton and Mike Brookes Imperial College London {rs8, mike.brookes}@imperial.ac.uk ABSTRACT Conventional beamformer design assumes that the phase differences

More information

High-speed Noise Cancellation with Microphone Array

High-speed Noise Cancellation with Microphone Array Noise Cancellation a Posteriori Probability, Maximum Criteria Independent Component Analysis High-speed Noise Cancellation with Microphone Array We propose the use of a microphone array based on independent

More information

Local Relative Transfer Function for Sound Source Localization

Local Relative Transfer Function for Sound Source Localization Local Relative Transfer Function for Sound Source Localization Xiaofei Li 1, Radu Horaud 1, Laurent Girin 1,2, Sharon Gannot 3 1 INRIA Grenoble Rhône-Alpes. {firstname.lastname@inria.fr} 2 GIPSA-Lab &

More information

Uplink and Downlink Beamforming for Fading Channels. Mats Bengtsson and Björn Ottersten

Uplink and Downlink Beamforming for Fading Channels. Mats Bengtsson and Björn Ottersten Uplink and Downlink Beamforming for Fading Channels Mats Bengtsson and Björn Ottersten 999-02-7 In Proceedings of 2nd IEEE Signal Processing Workshop on Signal Processing Advances in Wireless Communications,

More information

In air acoustic vector sensors for capturing and processing of speech signals

In air acoustic vector sensors for capturing and processing of speech signals University of Wollongong Research Online University of Wollongong Thesis Collection University of Wollongong Thesis Collections 2011 In air acoustic vector sensors for capturing and processing of speech

More information

Introduction to distributed speech enhancement algorithms for ad hoc microphone arrays and wireless acoustic sensor networks

Introduction to distributed speech enhancement algorithms for ad hoc microphone arrays and wireless acoustic sensor networks Introduction to distributed speech enhancement algorithms for ad hoc microphone arrays and wireless acoustic sensor networks Part I: Array Processing in Acoustic Environments Sharon Gannot 1 and Alexander

More information

Direction of Arrival Algorithms for Mobile User Detection

Direction of Arrival Algorithms for Mobile User Detection IJSRD ational Conference on Advances in Computing and Communications October 2016 Direction of Arrival Algorithms for Mobile User Detection Veerendra 1 Md. Bakhar 2 Kishan Singh 3 1,2,3 Department of lectronics

More information

Frequency Domain Analysis for Noise Suppression Using Spectral Processing Methods for Degraded Speech Signal in Speech Enhancement

Frequency Domain Analysis for Noise Suppression Using Spectral Processing Methods for Degraded Speech Signal in Speech Enhancement Frequency Domain Analysis for Noise Suppression Using Spectral Processing Methods for Degraded Speech Signal in Speech Enhancement 1 Zeeshan Hashmi Khateeb, 2 Gopalaiah 1,2 Department of Instrumentation

More information

Michael Brandstein Darren Ward (Eds.) Microphone Arrays. Signal Processing Techniques and Applications. With 149 Figures. Springer

Michael Brandstein Darren Ward (Eds.) Microphone Arrays. Signal Processing Techniques and Applications. With 149 Figures. Springer Michael Brandstein Darren Ward (Eds.) Microphone Arrays Signal Processing Techniques and Applications With 149 Figures Springer Contents Part I. Speech Enhancement 1 Constant Directivity Beamforming Darren

More information

A COHERENCE-BASED ALGORITHM FOR NOISE REDUCTION IN DUAL-MICROPHONE APPLICATIONS

A COHERENCE-BASED ALGORITHM FOR NOISE REDUCTION IN DUAL-MICROPHONE APPLICATIONS 18th European Signal Processing Conference (EUSIPCO-21) Aalborg, Denmark, August 23-27, 21 A COHERENCE-BASED ALGORITHM FOR NOISE REDUCTION IN DUAL-MICROPHONE APPLICATIONS Nima Yousefian, Kostas Kokkinakis

More information

SPEECH ENHANCEMENT BASED ON A LOG-SPECTRAL AMPLITUDE ESTIMATOR AND A POSTFILTER DERIVED FROM CLEAN SPEECH CODEBOOK

SPEECH ENHANCEMENT BASED ON A LOG-SPECTRAL AMPLITUDE ESTIMATOR AND A POSTFILTER DERIVED FROM CLEAN SPEECH CODEBOOK 18th European Signal Processing Conference (EUSIPCO-2010) Aalborg, Denmar, August 23-27, 2010 SPEECH ENHANCEMENT BASED ON A LOG-SPECTRAL AMPLITUDE ESTIMATOR AND A POSTFILTER DERIVED FROM CLEAN SPEECH CODEBOOK

More information

Airo Interantional Research Journal September, 2013 Volume II, ISSN:

Airo Interantional Research Journal September, 2013 Volume II, ISSN: Airo Interantional Research Journal September, 2013 Volume II, ISSN: 2320-3714 Name of author- Navin Kumar Research scholar Department of Electronics BR Ambedkar Bihar University Muzaffarpur ABSTRACT Direction

More information

BREAKING DOWN THE COCKTAIL PARTY: CAPTURING AND ISOLATING SOURCES IN A SOUNDSCAPE

BREAKING DOWN THE COCKTAIL PARTY: CAPTURING AND ISOLATING SOURCES IN A SOUNDSCAPE BREAKING DOWN THE COCKTAIL PARTY: CAPTURING AND ISOLATING SOURCES IN A SOUNDSCAPE Anastasios Alexandridis, Anthony Griffin, and Athanasios Mouchtaris FORTH-ICS, Heraklion, Crete, Greece, GR-70013 University

More information

Eigenvalues and Eigenvectors in Array Antennas. Optimization of Array Antennas for High Performance. Self-introduction

Eigenvalues and Eigenvectors in Array Antennas. Optimization of Array Antennas for High Performance. Self-introduction Short Course @ISAP2010 in MACAO Eigenvalues and Eigenvectors in Array Antennas Optimization of Array Antennas for High Performance Nobuyoshi Kikuma Nagoya Institute of Technology, Japan 1 Self-introduction

More information

STAP approach for DOA estimation using microphone arrays

STAP approach for DOA estimation using microphone arrays STAP approach for DOA estimation using microphone arrays Vera Behar a, Christo Kabakchiev b, Vladimir Kyovtorov c a Institute for Parallel Processing (IPP) Bulgarian Academy of Sciences (BAS), behar@bas.bg;

More information

Title. Author(s)Sugiyama, Akihiko; Kato, Masanori; Serizawa, Masahir. Issue Date Doc URL. Type. Note. File Information

Title. Author(s)Sugiyama, Akihiko; Kato, Masanori; Serizawa, Masahir. Issue Date Doc URL. Type. Note. File Information Title A Low-Distortion Noise Canceller with an SNR-Modifie Author(s)Sugiyama, Akihiko; Kato, Masanori; Serizawa, Masahir Proceedings : APSIPA ASC 9 : Asia-Pacific Signal Citationand Conference: -5 Issue

More information

Sound Source Localization using HRTF database

Sound Source Localization using HRTF database ICCAS June -, KINTEX, Gyeonggi-Do, Korea Sound Source Localization using HRTF database Sungmok Hwang*, Youngjin Park and Younsik Park * Center for Noise and Vibration Control, Dept. of Mech. Eng., KAIST,

More information

Subspace Noise Estimation and Gamma Distribution Based Microphone Array Post-filter Design

Subspace Noise Estimation and Gamma Distribution Based Microphone Array Post-filter Design Chinese Journal of Electronics Vol.0, No., Apr. 011 Subspace Noise Estimation and Gamma Distribution Based Microphone Array Post-filter Design CHENG Ning 1,,LIUWenju 3 and WANG Lan 1, (1.Shenzhen Institutes

More information

MMSE STSA Based Techniques for Single channel Speech Enhancement Application Simit Shah 1, Roma Patel 2

MMSE STSA Based Techniques for Single channel Speech Enhancement Application Simit Shah 1, Roma Patel 2 MMSE STSA Based Techniques for Single channel Speech Enhancement Application Simit Shah 1, Roma Patel 2 1 Electronics and Communication Department, Parul institute of engineering and technology, Vadodara,

More information

Speech Enhancement using Wiener filtering

Speech Enhancement using Wiener filtering Speech Enhancement using Wiener filtering S. Chirtmay and M. Tahernezhadi Department of Electrical Engineering Northern Illinois University DeKalb, IL 60115 ABSTRACT The problem of reducing the disturbing

More information

Speech enhancement with ad-hoc microphone array using single source activity

Speech enhancement with ad-hoc microphone array using single source activity Speech enhancement with ad-hoc microphone array using single source activity Ryutaro Sakanashi, Nobutaka Ono, Shigeki Miyabe, Takeshi Yamada and Shoji Makino Graduate School of Systems and Information

More information

Springer Topics in Signal Processing

Springer Topics in Signal Processing Springer Topics in Signal Processing Volume 3 Series Editors J. Benesty, Montreal, Québec, Canada W. Kellermann, Erlangen, Germany Springer Topics in Signal Processing Edited by J. Benesty and W. Kellermann

More information

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment International Journal of Electronics Engineering Research. ISSN 975-645 Volume 9, Number 4 (27) pp. 545-556 Research India Publications http://www.ripublication.com Study Of Sound Source Localization Using

More information

Speech Enhancement using Multiple Transducers

Speech Enhancement using Multiple Transducers Speech Enhancement using Multiple Transducers Craig Anderson A Thesis submitted to the Victoria University of Wellington in fulfilment of the requirements for the degree of Master of Engineering Victoria

More information

Indoor Localization based on Multipath Fingerprinting. Presented by: Evgeny Kupershtein Instructed by: Assoc. Prof. Israel Cohen and Dr.

Indoor Localization based on Multipath Fingerprinting. Presented by: Evgeny Kupershtein Instructed by: Assoc. Prof. Israel Cohen and Dr. Indoor Localization based on Multipath Fingerprinting Presented by: Evgeny Kupershtein Instructed by: Assoc. Prof. Israel Cohen and Dr. Mati Wax Research Background This research is based on the work that

More information

ENERGY-VS-PERFORMANCE TRADE-OFFS IN SPEECH ENHANCEMENT IN WIRELESS ACOUSTIC SENSOR NETWORKS

ENERGY-VS-PERFORMANCE TRADE-OFFS IN SPEECH ENHANCEMENT IN WIRELESS ACOUSTIC SENSOR NETWORKS ENERGY-VS-PERFORMANCE TRADE-OFFS IN SPEECH ENHANCEMENT IN WIRELESS ACOUSTIC SENSOR NETWORKS Fernando de la Hucha Arce 1, Fernando Rosas, Marc Moonen 1, Marian Verhelst, Alexander Bertrand 1 KU Leuven,

More information

Adaptive Systems Homework Assignment 3

Adaptive Systems Homework Assignment 3 Signal Processing and Speech Communication Lab Graz University of Technology Adaptive Systems Homework Assignment 3 The analytical part of your homework (your calculation sheets) as well as the MATLAB

More information

Speech Signal Enhancement Techniques

Speech Signal Enhancement Techniques Speech Signal Enhancement Techniques Chouki Zegar 1, Abdelhakim Dahimene 2 1,2 Institute of Electrical and Electronic Engineering, University of Boumerdes, Algeria inelectr@yahoo.fr, dahimenehakim@yahoo.fr

More information

Joint DOA and Array Manifold Estimation for a MIMO Array Using Two Calibrated Antennas

Joint DOA and Array Manifold Estimation for a MIMO Array Using Two Calibrated Antennas 1 Joint DOA and Array Manifold Estimation for a MIMO Array Using Two Calibrated Antennas Wei Zhang #, Wei Liu, Siliang Wu #, and Ju Wang # # Department of Information and Electronics Beijing Institute

More information

Different Approaches of Spectral Subtraction Method for Speech Enhancement

Different Approaches of Spectral Subtraction Method for Speech Enhancement ISSN 2249 5460 Available online at www.internationalejournals.com International ejournals International Journal of Mathematical Sciences, Technology and Humanities 95 (2013 1056 1062 Different Approaches

More information

ADAPTIVE ANTENNAS. TYPES OF BEAMFORMING

ADAPTIVE ANTENNAS. TYPES OF BEAMFORMING ADAPTIVE ANTENNAS TYPES OF BEAMFORMING 1 1- Outlines This chapter will introduce : Essential terminologies for beamforming; BF Demonstrating the function of the complex weights and how the phase and amplitude

More information

Noise Reduction: An Instructional Example

Noise Reduction: An Instructional Example Noise Reduction: An Instructional Example VOCAL Technologies LTD July 1st, 2012 Abstract A discussion on general structure of noise reduction algorithms along with an illustrative example are contained

More information

SPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN. Yu Wang and Mike Brookes

SPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN. Yu Wang and Mike Brookes SPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN Yu Wang and Mike Brookes Department of Electrical and Electronic Engineering, Exhibition Road, Imperial College London,

More information

Towards an intelligent binaural spee enhancement system by integrating me signal extraction. Author(s)Chau, Duc Thanh; Li, Junfeng; Akagi,

Towards an intelligent binaural spee enhancement system by integrating me signal extraction. Author(s)Chau, Duc Thanh; Li, Junfeng; Akagi, JAIST Reposi https://dspace.j Title Towards an intelligent binaural spee enhancement system by integrating me signal extraction Author(s)Chau, Duc Thanh; Li, Junfeng; Akagi, Citation 2011 International

More information