Linear and Parametric Microphone Array Processing
|
|
- Patrick Strickland
- 5 years ago
- Views:
Transcription
1 Part 5 - Joint Linear and Parametric Spatial Processing Emanuël A. P. Habets 1 and Sharon Gannot 2 1 International Audio Laboratories Erlangen, Germany A joint institution of University of Erlangen-Nuremberg and Fraunhofer IIS 2 Faculty of Engineering, Bar-Ilan University, Israel ICASSP 213, Vancouver, Canada
2 Overvie 1 Motivation 2 Informed Spatial Filtering 3 Examples Page 1/31
3 1. Motivation Classical Linear Spatial Filtering: + High amount of noise plus interference reduction + Controllable tradeoff beteen speech distortion and noise reduction + Controllable tradeoff beteen different noise types Not very robust.r.t. estimation errors, position changes, etc. Relatively slo response time Parametric Spatial Filtering: + Fast response time + Relatively robust.r.t. estimation errors, position changes, etc. + Possibility to manipulate parameters (e.g., virtual source displacement) Inherent tradeoff beteen speech distortion and noise reduction Model violations can introduce audible artifacts [Thiergart and Habets, 212] Relatively poor interference reduction due to tradeoff and model violations Page 2/31
4 Overvie 1 Motivation 2 Informed Spatial Filtering 3 Examples Page 3/31
5 2. Informed Spatial Filtering The main idea behind informed filtering is to incorporate relevant information about specific problem into design of filters and estimation of required statistics. Microphone Signals Informed Multichannel Spatial Filter Processed Signals Estimate Second Order Statistics Estimate Parameters (e.g., diffuseness, DOA) Figure: Informed filtering approach. Page 4/31
6 2. Informed Spatial Filtering A selection of parameters that can be used (see Part 4): Signal-to-diffuse ratio (SDR): Γ(k, m, p i) = P dir(k, m, p i) P diff (k, m), here P dir is poer of direct component at position p i and P diff is poer of diffuse component (assuming a ly homogenous sound field). Time and frequency dependent direction-of-arrival estimates. Time and frequency dependent interaural level differences. Time and frequency dependent interaural phase differences.... Page 5/31
7 Overvie 1 Motivation 2 Informed Spatial Filtering 3 Examples Example A: Extracting Coherent Sound Sources Example B: Dereverberation in SH Domain Example C: Directional Filtering Example D: Source Extraction Page 6/31
8 3.1 Example A: Extracting Coherent Sound Sources Signal model: y(k, m) = x(k, m) + v(k, m). Assumption: Desired signals are strongly coherent across array. Aim: Estimate X1(k, m) using a parametric multichannel Wiener filter [Benesty et al., 211]: Φ 1 v (k, m)φ x(k, m) h PMWF(k, m) = λ(k, m) + tr { Φ 1 v (k, m)φ x(k, m) } u1 5 λ Figure: Mapping from input signal-to-diffuse ratio to tradeoff parameter λ [Taseska and Habets, 212]. Page 7/31
9 Proposed Solution [Taseska and Habets, 212] y(k, m) Z(k, m) Parametric Multichannel Wiener Filter Estimate Noise PSD Matrix v(k, m) Estimate Speech Presence Probability P [H 1 y(k, m)] (k, m) Estimate Signal-to- Diffuse Ratio Figure: Block diagram of proposed system. Page 8/31
10 Algorithm Summary High-level description of proposed algorithm [Taseska and Habets, 212]: 1. Compute signal-to-diffuse ratio (SDR) using [Thiergart et al., 212]. 2. Compute a priori speech presence probability (SPP) based on SDR. 3. Compute multichannel a posteriori SPP [Souden et al., 21]. 4. Update noise PSD matrix using a posteriori SPP. 5. Compute tradeoff parameter for parametric multichannel Wiener filter (PMWF) based on SDR: - When SDR is high, e decrease amount of speech distortion. - When SDR is lo, e increase amount of noise reduction. 6. Compute and apply parametric multichannel Wiener filter. Page 9/31
11 Results (1).8 λ = λ = 1 λ = f(γ) λ = f(γ) + SPP.8 Δ PESQ input SNR (db) input SNR (db) Figure: Performance evaluation: PESQ improvement for stationary diffuse noise (left) and diffuse babble speech (right) [Taseska and Habets, 212]. Page 1/31
12 Results (2) λ = λ = 1 λ = f(γ) λ = f(γ) + SPP SNR gain (db) input SNR (db) input SNR (db) Figure: Performance evaluation: segmental SNR improvement for stationary diffuse noise (left) and diffuse babble speech (right) [Taseska and Habets, 212]. Page 11/31
13 Results (3) Frequency [khz] Frequency [khz] Time [s] Time [s] 5 (a) First Microphone Signal (b) MVDR Frequency [khz] Frequency [khz] Time [s] (c) Parametric MWF Time [s] (d) Parametric MWF ith MC-SPP 5 Figure: Examples obtained using M=4 microphone signals corrupted by sensor noise and babble speech (input SNR = 1 db). Audio Examples Page 12/31
14 3.2 Example B: Dereverberation in SH Domain Assumed signal model ith stacked spherical harmonic components: p(k, m) = x(k, m) + d(k, m) + ṽ(k, m) }{{} = γ(k, m) X (k, m) + ũ(k, m) γ(k, m) = x(k, m) X (k, m) = y(ω dir) Y (Ω dir ) = γ dir, here Y is zero-order spherical harmonic and Ω dir is DOA. p d v x Spherical Harmonics up to order 3 Page 13/31
15 Proposed Solution [Braun et al., 213] Desired signal: The direct signal component X(k, m) hich corresponds to sound pressure measured at center of array in absence of spherical microphone array. Assumption: Direct, diffuse and noise components are mutually uncorrelated. Proposed solution: The (rank-1) MWF provides an MMSE estimate of X (k, m). For practical reasons, e split MWF into an MVDR filter folloed by a single-channel Wiener filter: h MWF(k, m) = φ X (k, m) Φ 1 ũ (k, m) γ dir φ X (k, m) γ H dir Φ 1 ũ (k, m) γ dir + 1 φ X = Φ 1 ũ (k, m) γ dir γ H dir Φ 1 ũ (k, m) γ dir φ }{{} X + [ γ H dir Φ 1 ũ (k, m) γ ] 1 dir }{{} h MVDR (k,m) H W (k,m) Page 14/31
16 Parameter-based PSD Matrix Estimation Required information: Diffuse PSD matrix estimation: Direction of arrival (DOA) γ dir Interference PSD matrix: Φũ(k, m) = Φ d(k, m) + Φṽ(k, m) Assume model for diffuse sound component: Φ d(k, m) = φ D (k, m) I (L+1) 2 Calculate diffuse sound PSD using an estimate of diffuseness Ψ: φ D (k, m) = φ P (k, m) φṽ (k, m) Ψ 1 (k, m) STFT SHT STFT -1 Diffuseness Estimation Residual Interf. PSD Diffuse PSD Estimation Page 15/31
17 Results Frequency [khz] Frequency [khz] Time [s] Time [s] 5 (a) Reference X (k, m) (b) Received P (k, m) Frequency [khz] Frequency [khz] Time [s] Time [s] 5 (c) Processed: MVDR (d) Processed: MWF Figure: Examples obtained using simulated signals [Jarrett et al., 212] (source-array distance is 2 m, SNR = 2 db, T 6 =4 ms). Audio examples. Page 16/31
18 3.3 Example C: Directional Filtering Flexible sound acquisition in noisy and reverberant environments ith rapidly changing acoustic scenes is a common problem in modern communication systems. A filter is proposed that provides an arbitrary response for J sources being simultaneously active per time and frequency. The proposed filter provides an optimal tradeoff beteen hite noise gain (WNG) and directivity index. The filter exploits instantaneous information on sound (narroband DOAs, diffuse-to-noise ratio) hich allos a nearly immediate adaption to changes in acoustic scene. Page 17/31
19 Problem Formulation Assuming three components in (1) are mutually uncorrelated, e canthe expressdesired poer signal spectral density is given (PSD) by: matrix of microphone signals as { J } Φ(k, n) Z(k, =E m) x(k, n) = x H (k, n) G(k, ϕ j)x (j) 1 L j=1 = Φl(k, n)+φd(k, n)+φn(k, n), (2) ith Signal model: Based on a multi-ave sound field model, M microphone signals can be expressed as: J y(k, m) = x (j) (k, m) + d(k, m) + v(k, m) }{{}}{{} j=1 }{{} diffuse sound sensor noise J plane aves Aim: Capturing J plane aves (J M) ith desired arbitrary gain hile attenuating sensor noise and reverberation. l=1 (k, m) Φd(k, n) =φd(k, n) Γd(k), (3) Φn(k, n) =φn(k, n) I. (4) Here, I is an identity matrix, φn(k, n) is expected poer of microphone self-noise, hich is identical for all microphones, and φd(k, Linear n) isand Parametric expected poer Microphone of Array diffuse Processing field, hich can vary rapidly Emanuël across Habets time and (FAU) frequency. and Sharon The Gannot ij-th(biu) element of coher- G(k, ϕ) 2 [db] G1 G2 ϕa ϕb DOA ϕ [ ] The desired signal is estimated using an informed LCMV filter: Ẑ(k, m) = h H ilcmv(k, m) y(k, m) 1. Toarbitrarydirectivityfunctions&sourcepositions 3.1. Existing Spatial Filters 3. OPTIMAL SPATIAL FILTERS While PSD φn(k, c International n) can beaudio estimated Laboratories during Erlangen, periods of213 silence, φd(k, n) is commonly assumed unknon and unobservable. Page 18/31 We
20 Proposed Solution (1) The proposed informed LCMV filer is given by: h ilcmv = argmin h h H [Φ d (k, m) + Φ v(k, m)] h s. t. h H (k, m) a(k, ϕ j) = G(k, ϕ j), j {1, 2,..., J} here a(k, ϕ j) denotes steering vector for jth plane ave at time m and frequency k. For assumed signal model, e can alternatively minimize h H [Ψ(k, m) Γ d (k) + I] h, here Ψ(k, m) denotes instantaneous diffuse-to-noise ratio (DNR) and Γ d (k) denotes coherence matrix of diffuse sound field. The filter is updated for each time and frequency given instantaneous parametric information (DOAs, DNR). The filter requires knoledge of DNR, hich can be estimated using an auxiliary filter (see poster session AASP-P8 on Friday or [Thiergart and Habets, 213]). Page 19/31
21 Proposed Solution (2) DI [db] WNG [db] n 5d 4nd nd 6 3 frequency [khz] DNR Estimation DOA Estimation Desired Response 9 45 Filter Weights frequency time [khz] [s] time [s] (b) Mean (a) DOA WNG ϕ1(k, n) [ ] (b) G(k, ϕ1) 2 [db] 3. DI and WNG of filters in Sec. 3. For d, Figure: Left: DOA ϕ minimum WNG 5 1 (k, m) as a function of time 4. EstimatedDOAϕ1(k, and frequency. n) Right: and resulting Desired gains G(k, ϕ1) as set to 12 db to make filter robust response G(k, ϕ 5 against microphone 4 self-noise. 1 ) 2 in db for DOA ϕ 1 (k, m) as a function of time and frequency. fit. The best performance in terms of SSNR is obtained using n. 3 1 In terms of PESQ, nd and d outperform n. Usinginstanta- neous directional c constraints International (asaudio in this Laboratories section) instead Erlangen, of 213 time- for alinear silent part and 2 Parametric of signal Microphone and duringarray speech Processing activity [both signal 15 partsemanuël marked in 2(b)]. During silence, proposed filter 1 Habets (FAU) and Sharon Gannot (BIU) invariant constrains (as in Sec. 5.1, values in brackets) Page mainly 2/31 re- frequency [khz] n d nd nd 1 frequency [khz] (a) Mean DI frequency [khz] frequency [khz] time [s] (a) DOA ϕ1(k, n) [ ]
22 Informed InformedLCMV LCMVFilter Filter Results (1) DIDI[dB] [db] DIDI[dB] [db].(9). t. (9). (12) (12) pressed essed as as nn dn dn dnd dnd nd nd nd nd ndnd sources active sources active sources silent sources silent time time [s] [s] True Ψ(k, [db] (a)(a) True Ψ(k, n) n) [db] nes. The solu. The solu frequency [khz] frequency [khz] (14) (14) (15) (15) WNG WNG[dB] [db] WNG WNG[dB] [db] t.9),(9), (13) (13) frequency [khz] frequency [khz] frequency [khz] frequency [khz] Mean (a)(a) Mean DIDI Mean (a)(a) Mean DIDI nn dn dn nd nd d d nd nd nd nd ndnd sources silent sources silent sources active sources active frequency [khz] frequency [khz] frequency frequency[khz] [khz] frequency frequency[khz] [khz] 7 7 frequency [khz] frequency [khz] time time [s] [s] he propagation propagation Mean WNG (b)(b) Mean WNG Estimated Ψ(k, [db] (b)(b) Estimated Ψ(k, n) n) [db] Mean WNG (b)(b) Mean WNG ns given areare given byby WNG filters Sec., 3. 3.DIDI andand WNG of of filters in in Sec ForFor d, d nf of Ψ(k, Ψ(k, n)n) is is 3.DIDI and WNG of filters Sec., True estimated DNR Ψ(k, The to marked areas 3. and WNG of filters in in Sec ForFor d, robust 2. 2.True andand estimated DNR Ψ(k, n).n).the to marked areas d minimum WNG as 12 make filter minimum WNG as setset to to 12 dbdb to to make filter robust Figure: Top: Directivity index (DI) in db. Figure: Top: True DNR in db. Bottom: indicate respectively a silent and active part signal. indicate respectively a silent and active part of of signal. minimum WNG as set to 12 db to make filter robust minimum WNG as set to 12 db to make filter robust against microphone self-noise. against microphone self-noise. against microphone self-noise. against microphone self-noise. Bottom: White noise gain (WNG) in db. Estimated DNR in db. frequency frequency[khz] [khz] frequency frequency[khz] [khz] frequency [khz] frequency [khz] lem m in in (8)(8) areare er diffuse andand diffuse Es Estim Es Estim Table 1 Table 1 su Table 1 Table 1 su International Erlangen International AudioLabs Erlangen terms in in terms of of sigs The required expected poer microphone self-noise n) minimizes The required expected poer of of microphone self-noise φn φ (k, n)audiolabs n (k, noise poer, The in terms nactivity terms of of sigs speech activity reverberant environment. The estimated in destimated speech duedue to to reverberant environment. ratio (SRR) ratio (SRR), a example estimated during silence assuming that poer cancan forfor example be be estimated during silence assuming that poer 6 activity 6speech activity due to reverberant environment. The estimated speech due to reverberant environment. The estimated ratio (SRR) (SRR), DNR 2(b) possesses a limited temporal resolution due to DNR in in 2(b) possesses a limited temporal resolution due to ratio ed approach approach in in minimizes diffuse poer, is SRR (sourc SRR (source a nd duedue is constant over time.note Note that proposed DNR estimator does is constant over time. that proposed DNR estimator does DNR 2(b) possesses a limited temporal resolution DNR in in 2(b) possesses a limited temporal resolution to to SRR (sourc (source an obtained incorporated temporal averaging process. Neverless, Ψ(k, be be obtained incorporated temporal averaging process. Neverless, Ψ(k, n)n) SRR ( ) and th ( ) and necessarily provide loest estimation variance practice nd notnot necessarily provide loest estimation variance in in practice nd incorporated temporal averaging process. Neverless, Ψ(k, n) incorporated temporal averaging process. Neverless, Ψ(k, n) proposed ilcmv filter that minimizes (( ) and th ssprit ESPRIT [24] [24] estimates sufficiently accurate shon folloing results. estimates areare sufficiently accurate as as shon byby folloing results. nd )ndand formance. HoH chosen optimization criteria (16), provides unbiased duedue to to chosen optimization criteria (16), butbut provides unbiased estimates sufficiently accurate shon folloing results. formance. estimates areare sufficiently accurate as as shon byby folloing results. formance.hoh estimation stimation of of Figure 3(a) depicts mean and both formance. Figure 3(a) depicts mean DIDI forfor n[shon d(hich areare both diffuse plus noise poer hen n and d (hich than SS SSNR results. results. Figure 3(a) depicts mean DI (hich are both than Figure 3(a) depicts for mean DIproposed forfor n and dfilter both n and d (hich than SS than SSNR signal-independent), for filter are (hich signal-independent), andand proposed nd (hich nd The bes fit.fit.the best p sources are active (red solid line) and signal-independent), and for proposed filter (hich signal-independent), and for proposed filter (hich nd nd onal fit.the The besp l fil-filfit.in best is signal-dependent). proposed filter, sho is signal-dependent). ForFor proposed filter, ee sho DIDI In terms of terms of P is signal-dependent). proposed filter, sho is signal-dependent). ForFor proposed filter, ee sho DIDI In In EXPERIMENTAL RESULTS 5. 5.EXPERIMENTAL RESULTS ffuse sound use sound is is terms of terms of P silent (red dashed line)]. a silent part signal during speech activity [both signal forfor a silent part of of signal andand during speech activity [both signal neousdirectio direc neous for a silent part of signal and during speech activity [both signal for a silent part of signal and during speech activity [both signal by maximizing maximizing neousdirectio direc neous parts marked 2(b)]. During silence, proposed filter invariant parts marked in in 2(b)]. During silence, proposed filter invariant co cons assume L= 2 plane aves model ULA LetLet us us assume L= 2 plane aves in in model in in (1)(1) andand an an ULA parts marked in 2(b)]. During silence, proposed filter invariant parts marked in During silence, filter invariant co cons (dashed line ) 2(b)]. provides same lo DI as. During speech (dashed line same lo DI asproposed n. nduring speech nd )ndprovides duced duced acha ith M 4 microphones ith an inter-microphone spacing 3 cm. ith MParametric = 4=microphones ith anarray inter-microphone of of 3 cm. (dashed line )ndprovides ) provides same lo as. During speech213 (dashed line nd same lo DIisDI as. nduring speech c), International Linear and Microphone Processing spacing Audio Laboratories Erlangen, n duced duced acha activity (solid line ), obtained DI as high as for robust activity (solid line obtained DI is as high as for robust 3 3 nd nd varying sou varying source reverberant shoebox room (7., RT 38 ms) AA reverberant shoebox room ( mm, RT 38 ms) 6 6 activity (solid line ),nd ), obtained DI is as high as for robust (16) activity (solid line obtained DI is as high as for robust (16) nd varying sou varying source Emanue l Habets (FAU) and Sharon Gannot (BIU) Page 21/31 beamformer ). Figure 3(b) shos corresponding WNGs. quired SDSD beamformer (( 3(b) shos corresponding WNGs. d ).dfigure quired comp comple as simulated using source-image method [26, ith to as simulated using source-image method [26, 27]27] ith to beamformer ). Figure 3(b) shos corresponding WNGs. SDSD beamformer (( ). Figure 3(b) shos corresponding WNGs.
23 Results (2) The proposed filter provides a high DI hen sound field is diffuse and a high WNG hen sensor noise is dominant. Interfering sound can be strongly attenuated if desired. The proposed DNR estimator provides a sufficiently high accuracy and temporal resolution to allo signal enhancement under adverse conditions even in changing acoustic scenes. SegSIR [db] SegSRR [db] SegSNR [db] PESQ 11 (11) 7 ( 7) 26 (26) 1.5 (1.5) n 21 (32) 2 ( 3) 33 (31) 2. (1.7) d 26 (35) ( 1) 22 (24) 2.1 (2.) nd 25 (35) 1 ( 1) 28 (26) 2.1 (2.) Table: Performance of all filters [ unprocessed, first sub-column using true DOAs (of sources), second sub-column using estimated DOAs (of plane aves)]. Audio Examples Page 22/31
24 3.4 Example D: Source Extraction Scenario Multiple talkers Additive background noise Distributed sensor arrays Applications Spatial filter Estimate of desired source at microphone Teleconferencing systems Automatic speech recognition Spatial sound reproduction Signal model: y(k, m) = x (d) (k, m) + i d x (i) (k, m) + v(k, m). (d) Aim: Obtain an MMSE estimate of X 1 (k, m). Page 23/31
25 Proposed Solution [Taseska and Habets, 213] Hyposes: H v : y(k, m) = v(k, m) speech absent H x : y(k, m) = x(k, m) + v(k, m) speech present J Hx j : y(k, m) = x (j) (k, m) + x (i) (k, m) + v(k, m) j = 1, 2,..., J i j } {{ } Recursive estimation of PSD matrices: ( Φ (j) x (m) = p[hx j y] α Φ(j) x x (m 1) + (1 α x) yy H) ( + 1 p[hx j y] ) Φ(j) x (m 1) Signal-to-diffuse ratio (Γ) and position (Θ) -based posterior probabilities: p[h j x y] = p[h j x y, H x] p[h x y] p[h j x Θ, H x] p[h x Γ, y] Page 24/31
26 Parameter-based PSD Matrix Estimation Γ p[h x Γ, y] Φv Ω Θ p[h j x Θ, H x] ˆp[Hj x y] Φ (j) x The distribution p[ Θ Hx] is modelled as a Gaussian mixture (GM). GM parameters estimated by Expectation-Maximization algorithm. Page 25/31
27 Results (1) Setup: Three reverberant sources ith approximately equal poer, diffuse babble speech (SNR=22 db), and uncorrelated sensor noise (SNR =5 db). The reverberation time as T6 = 25 ms. To uniform circular arrays ere used ith three omnidirectional microphones, a diameter 2.5 cm and an inter-array spacing of 1.5 m. (a) Training during single-talk (b) Training during triple-talk Figure: Output of EM algorithm (3 iterations) and 4.5 s of noisy speech data. The actual source positions are denoted by hite squares. The array location is marked by a plus symbol. The interior of each ellipse contains 85% probability mass of respective Gaussian. Page 26/31
28 Results (2) mixture mixture reference source signals reference source signals (1) (2) (3) extracted source signals extracted source signals (1) (2) (3) time (s) time (s) Figure: Left: constant triple-talk scenario. Right: mainly single-talk scenario. (S,M) Audio files available at Page 27/31
29 More Information These and or examples are presented at ICASSP 213 on: Friday 1:3-12:3 in Poster Session AASP-P8: An Informed Spatial Filter in Spherical Harmonic Domain for Joint Noise Reduction and Dereverberation (Braun, Jarret, Fischer and Habets) Friday 1:3-12:3 in Poster Session AASP-P8: An Informed LCMV Filter Based on Multiple Instantaneous Direction-Of-Arrival Estimates (Thiergart and Habets) Friday 1:3-12:3 in Poster Session AASP-P8 MMSE-based Source Extraction using Position-based Posterior Probabilities (Taseska and Habets) Friday 1:3-12:3 in Poster Session AASP-P8 Spherical Harmonic Domain Noise Reduction Using an MVDR Beamformer and DOA-based Second-order Statistics Estimation (Jarrett, Habets and Naylor) Page 28/31
30 Special thanks to Sebastian Braun, Maja Taseska, Oliver Thiergart and Daniel Jarrett for ir contributions. Page 29/31
31 References I Benesty, J., Chen, J., and Habets, E. A. P. (211). Speech Enhancement in STFT Domain. SpringerBriefs in Electrical and Computer Engineering. Springer-Verlag. Braun, S., Jarrett, D. P., Fischer, J., and Habets, E. A. P. (213). An informed filter for dereverberation in spherical harmonic domain. In Proc. IEEE Intl. Conf. on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, Canada. Jarrett, D. P., Habets, E. A. P., Thomas, M. R. P., and Naylor, P. A. (212). Rigid sphere room impulse response simulation: algorithm and applications. J. Acoust. Soc. Am., 132(3): Souden, M., Chen, J., Benesty, J., and Affes, S. (21). Gaussian model-based multichannel speech presence probability. IEEE Trans. Audio, Speech, Lang. Process., 18(5): Taseska, M. and Habets, E. (213). MMSE-based source extraction using position-based posterior probabilities. In Proc. IEEE Intl. Conf. on Acoustics, Speech and Signal Processing (ICASSP). Page 3/31
32 References II Taseska, M. and Habets, E. A. P. (212). MMSE-based blind source extraction in diffuse noise fields using a complex coherence-based a priori SAP estimator. In Proc. Intl. Workshop Acoust. Signal Enhancement (IWAENC). Thiergart, O., Del Galdo, G., and Habets, E. A. P. (212). On coherence in mixed sound fields and its application to signal-to-diffuse ratio estimation. J. Acoust. Soc. Am., 132(4): Thiergart, O. and Habets, E. (213). Informed optimum filtering using multiple instantaneous direction-of-arrival estimates. In Proc. IEEE Intl. Conf. on Acoustics, Speech and Signal Processing (ICASSP). Thiergart, O. and Habets, E. A. P. (212). Sound field model violations in parametric sound processing. In Proc. Intl. Workshop Acoust. Signal Enhancement (IWAENC). Page 31/31
Recent Advances in Acoustic Signal Extraction and Dereverberation
Recent Advances in Acoustic Signal Extraction and Dereverberation Emanuël Habets Erlangen Colloquium 2016 Scenario Spatial Filtering Estimated Desired Signal Undesired sound components: Sensor noise Competing
More informationEmanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor. Presented by Amir Kiperwas
Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor Presented by Amir Kiperwas 1 M-element microphone array One desired source One undesired source Ambient noise field Signals: Broadband Mutually
More informationInformed Spatial Filtering for Sound Extraction Using Distributed Microphone Arrays
IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 7, JULY 2014 1195 Informed Spatial Filtering for Sound Extraction Using Distributed Microphone Arrays Maja Taseska, Student
More informationarxiv: v1 [cs.sd] 4 Dec 2018
LOCALIZATION AND TRACKING OF AN ACOUSTIC SOURCE USING A DIAGONAL UNLOADING BEAMFORMING AND A KALMAN FILTER Daniele Salvati, Carlo Drioli, Gian Luca Foresti Department of Mathematics, Computer Science and
More informationEUSIPCO
EUSIPCO 97 AN INFORMED MMSE FILTER BASED ON MULTIPLE INSTANTANEOUS DIRECTION-OF-ARRIVAL ESTIMATES Oliver Thiergart, Maja Taseska, and Emanuël A. P. Habets International Audio Laboratories Erlangen Am Wolfsmantel,
More informationBlind Beamforming for Cyclostationary Signals
Course Page 1 of 12 Submission date: 13 th December, Blind Beamforming for Cyclostationary Signals Preeti Nagvanshi Aditya Jagannatham UCSD ECE Department 9500 Gilman Drive, La Jolla, CA 92093 Course Project
More informationIEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 5, MAY
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 5, MAY 2013 945 A Two-Stage Beamforming Approach for Noise Reduction Dereverberation Emanuël A. P. Habets, Senior Member, IEEE,
More informationSpeech Enhancement Using Microphone Arrays
Friedrich-Alexander-Universität Erlangen-Nürnberg Lab Course Speech Enhancement Using Microphone Arrays International Audio Laboratories Erlangen Prof. Dr. ir. Emanuël A. P. Habets Friedrich-Alexander
More informationA BROADBAND BEAMFORMER USING CONTROLLABLE CONSTRAINTS AND MINIMUM VARIANCE
A BROADBAND BEAMFORMER USING CONTROLLABLE CONSTRAINTS AND MINIMUM VARIANCE Sam Karimian-Azari, Jacob Benesty,, Jesper Rindom Jensen, and Mads Græsbøll Christensen Audio Analysis Lab, AD:MT, Aalborg University,
More informationMicrophone Array Design and Beamforming
Microphone Array Design and Beamforming Heinrich Löllmann Multimedia Communications and Signal Processing heinrich.loellmann@fau.de with contributions from Vladi Tourbabin and Hendrik Barfuss EUSIPCO Tutorial
More informationROBUST SUPERDIRECTIVE BEAMFORMER WITH OPTIMAL REGULARIZATION
ROBUST SUPERDIRECTIVE BEAMFORMER WITH OPTIMAL REGULARIZATION Aviva Atkins, Yuval Ben-Hur, Israel Cohen Department of Electrical Engineering Technion - Israel Institute of Technology Technion City, Haifa
More informationA MULTI-CHANNEL POSTFILTER BASED ON THE DIFFUSE NOISE SOUND FIELD. Lukas Pfeifenberger 1 and Franz Pernkopf 1
A MULTI-CHANNEL POSTFILTER BASED ON THE DIFFUSE NOISE SOUND FIELD Lukas Pfeifenberger 1 and Franz Pernkopf 1 1 Signal Processing and Speech Communication Laboratory Graz University of Technology, Graz,
More informationA Three-Microphone Adaptive Noise Canceller for Minimizing Reverberation and Signal Distortion
American Journal of Applied Sciences 5 (4): 30-37, 008 ISSN 1546-939 008 Science Publications A Three-Microphone Adaptive Noise Canceller for Minimizing Reverberation and Signal Distortion Zayed M. Ramadan
More informationJoint recognition and direction-of-arrival estimation of simultaneous meetingroom acoustic events
INTERSPEECH 2013 Joint recognition and direction-of-arrival estimation of simultaneous meetingroom acoustic events Rupayan Chakraborty and Climent Nadeu TALP Research Centre, Department of Signal Theory
More informationDual-Microphone Speech Dereverberation in a Noisy Environment
Dual-Microphone Speech Dereverberation in a Noisy Environment Emanuël A. P. Habets Dept. of Electrical Engineering Technische Universiteit Eindhoven Eindhoven, The Netherlands Email: e.a.p.habets@tue.nl
More informationMicrophone Array Power Ratio for Speech Quality Assessment in Noisy Reverberant Environments 1
for Speech Quality Assessment in Noisy Reverberant Environments 1 Prof. Israel Cohen Department of Electrical Engineering Technion - Israel Institute of Technology Technion City, Haifa 3200003, Israel
More informationEnhancement of Speech Signal Based on Improved Minima Controlled Recursive Averaging and Independent Component Analysis
Enhancement of Speech Signal Based on Improved Minima Controlled Recursive Averaging and Independent Component Analysis Mohini Avatade & S.L. Sahare Electronics & Telecommunication Department, Cummins
More informationDual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation
Dual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation Gal Reuven Under supervision of Sharon Gannot 1 and Israel Cohen 2 1 School of Engineering, Bar-Ilan University,
More informationSpeech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.
www.ijecs.in International Journal Of Engineering And Computer Science ISSN:2319-7242 Volume 4 Issue 4 April 2015, Page No. 11143-11147 Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya
More informationNOISE POWER SPECTRAL DENSITY MATRIX ESTIMATION BASED ON MODIFIED IMCRA. Qipeng Gong, Benoit Champagne and Peter Kabal
NOISE POWER SPECTRAL DENSITY MATRIX ESTIMATION BASED ON MODIFIED IMCRA Qipeng Gong, Benoit Champagne and Peter Kabal Department of Electrical & Computer Engineering, McGill University 3480 University St.,
More informationMULTICHANNEL AUDIO DATABASE IN VARIOUS ACOUSTIC ENVIRONMENTS
MULTICHANNEL AUDIO DATABASE IN VARIOUS ACOUSTIC ENVIRONMENTS Elior Hadad 1, Florian Heese, Peter Vary, and Sharon Gannot 1 1 Faculty of Engineering, Bar-Ilan University, Ramat-Gan, Israel Institute of
More informationOPTIMUM POST-FILTER ESTIMATION FOR NOISE REDUCTION IN MULTICHANNEL SPEECH PROCESSING
14th European Signal Processing Conference (EUSIPCO 6), Florence, Italy, September 4-8, 6, copyright by EURASIP OPTIMUM POST-FILTER ESTIMATION FOR NOISE REDUCTION IN MULTICHANNEL SPEECH PROCESSING Stamatis
More informationFlexible and efficient spatial sound acquisition and subsequent. Parametric Spatial Sound Processing
[ Konrad Kowalczyk, Oliver Thiergart, Maja Taseska, Giovanni Del Galdo, Ville Pulkki, and Emanuël A.P. Habets ] Parametric Spatial Sound Processing ear photo istockphoto.com/xrender assisted listening
More information260 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 2, FEBRUARY /$ IEEE
260 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 2, FEBRUARY 2010 On Optimal Frequency-Domain Multichannel Linear Filtering for Noise Reduction Mehrez Souden, Student Member,
More informationThe Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals
The Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals Maria G. Jafari and Mark D. Plumbley Centre for Digital Music, Queen Mary University of London, UK maria.jafari@elec.qmul.ac.uk,
More informationIEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 24, NO. 7, JULY
IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 24, NO. 7, JULY 2016 1291 Spotforming: Spatial Filtering With Distributed Arrays for Position-Selective Sound Acquisition Maja Taseska,
More informationPerformance Evaluation of Nonlinear Speech Enhancement Based on Virtual Increase of Channels in Reverberant Environments
Performance Evaluation of Nonlinear Speech Enhancement Based on Virtual Increase of Channels in Reverberant Environments Kouei Yamaoka, Shoji Makino, Nobutaka Ono, and Takeshi Yamada University of Tsukuba,
More informationHUMAN speech is frequently encountered in several
1948 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 20, NO. 7, SEPTEMBER 2012 Enhancement of Single-Channel Periodic Signals in the Time-Domain Jesper Rindom Jensen, Student Member,
More informationPerformance Evaluation of Capon and Caponlike Algorithm for Direction of Arrival Estimation
Performance Evaluation of Capon and Caponlike Algorithm for Direction of Arrival Estimation M H Bhede SCOE, Pune, D G Ganage SCOE, Pune, Maharashtra, India S A Wagh SITS, Narhe, Pune, India Abstract: Wireless
More informationIN REVERBERANT and noisy environments, multi-channel
684 IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 11, NO. 6, NOVEMBER 2003 Analysis of Two-Channel Generalized Sidelobe Canceller (GSC) With Post-Filtering Israel Cohen, Senior Member, IEEE Abstract
More information/$ IEEE
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 6, AUGUST 2009 1071 Multichannel Eigenspace Beamforming in a Reverberant Noisy Environment With Multiple Interfering Speech Signals
More informationMultiple Sound Sources Localization Using Energetic Analysis Method
VOL.3, NO.4, DECEMBER 1 Multiple Sound Sources Localization Using Energetic Analysis Method Hasan Khaddour, Jiří Schimmel Department of Telecommunications FEEC, Brno University of Technology Purkyňova
More informationAdaptive Beamforming Applied for Signals Estimated with MUSIC Algorithm
Buletinul Ştiinţific al Universităţii "Politehnica" din Timişoara Seria ELECTRONICĂ şi TELECOMUNICAŢII TRANSACTIONS on ELECTRONICS and COMMUNICATIONS Tom 57(71), Fascicola 2, 2012 Adaptive Beamforming
More informationDISTANT or hands-free audio acquisition is required in
158 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 1, JANUARY 2010 New Insights Into the MVDR Beamformer in Room Acoustics E. A. P. Habets, Member, IEEE, J. Benesty, Senior Member,
More informationAN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION
AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION Gerhard Doblinger Institute of Communications and Radio-Frequency Engineering Vienna University of Technology Gusshausstr. 5/39,
More informationAN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION
1th European Signal Processing Conference (EUSIPCO ), Florence, Italy, September -,, copyright by EURASIP AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION Gerhard Doblinger Institute
More informationSPECTRAL COMBINING FOR MICROPHONE DIVERSITY SYSTEMS
17th European Signal Processing Conference (EUSIPCO 29) Glasgow, Scotland, August 24-28, 29 SPECTRAL COMBINING FOR MICROPHONE DIVERSITY SYSTEMS Jürgen Freudenberger, Sebastian Stenzel, Benjamin Venditti
More informationBEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR
BeBeC-2016-S9 BEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR Clemens Nau Daimler AG Béla-Barényi-Straße 1, 71063 Sindelfingen, Germany ABSTRACT Physically the conventional beamforming method
More informationAntennas and Propagation. Chapter 5c: Array Signal Processing and Parametric Estimation Techniques
Antennas and Propagation : Array Signal Processing and Parametric Estimation Techniques Introduction Time-domain Signal Processing Fourier spectral analysis Identify important frequency-content of signal
More informationJoint dereverberation and residual echo suppression of speech signals in noisy environments Habets, E.A.P.; Gannot, S.; Cohen, I.; Sommen, P.C.W.
Joint dereverberation and residual echo suppression of speech signals in noisy environments Habets, E.A.P.; Gannot, S.; Cohen, I.; Sommen, P.C.W. Published in: IEEE Transactions on Audio, Speech, and Language
More informationSingle channel noise reduction
Single channel noise reduction Basics and processing used for ETSI STF 94 ETSI Workshop on Speech and Noise in Wideband Communication Claude Marro France Telecom ETSI 007. All rights reserved Outline Scope
More informationWIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY
INTER-NOISE 216 WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY Shumpei SAKAI 1 ; Tetsuro MURAKAMI 2 ; Naoto SAKATA 3 ; Hirohumi NAKAJIMA 4 ; Kazuhiro NAKADAI
More informationMULTICHANNEL systems are often used for
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 52, NO. 5, MAY 2004 1149 Multichannel Post-Filtering in Nonstationary Noise Environments Israel Cohen, Senior Member, IEEE Abstract In this paper, we present
More informationAdaptive Beamforming. Chapter Signal Steering Vectors
Chapter 13 Adaptive Beamforming We have already considered deterministic beamformers for such applications as pencil beam arrays and arrays with controlled sidelobes. Beamformers can also be developed
More informationONE of the most common and robust beamforming algorithms
TECHNICAL NOTE 1 Beamforming algorithms - beamformers Jørgen Grythe, Norsonic AS, Oslo, Norway Abstract Beamforming is the name given to a wide variety of array processing algorithms that focus or steer
More information546 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 4, MAY /$ IEEE
546 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL 17, NO 4, MAY 2009 Relative Transfer Function Identification Using Convolutive Transfer Function Approximation Ronen Talmon, Israel
More informationSpeech Enhancement for Nonstationary Noise Environments
Signal & Image Processing : An International Journal (SIPIJ) Vol., No.4, December Speech Enhancement for Nonstationary Noise Environments Sandhya Hawaldar and Manasi Dixit Department of Electronics, KIT
More informationDesign of Robust Differential Microphone Arrays
IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 10, OCTOBER 2014 1455 Design of Robust Differential Microphone Arrays Liheng Zhao, Jacob Benesty, Jingdong Chen, Senior Member,
More informationEffective post-processing for single-channel frequency-domain speech enhancement Weifeng Li a
R E S E A R C H R E P O R T I D I A P Effective post-processing for single-channel frequency-domain speech enhancement Weifeng Li a IDIAP RR 7-7 January 8 submitted for publication a IDIAP Research Institute,
More informationDirection-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method
Direction-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method Udo Klein, Member, IEEE, and TrInh Qu6c VO School of Electrical Engineering, International University,
More informationLETTER Pre-Filtering Algorithm for Dual-Microphone Generalized Sidelobe Canceller Using General Transfer Function
IEICE TRANS. INF. & SYST., VOL.E97 D, NO.9 SEPTEMBER 2014 2533 LETTER Pre-Filtering Algorithm for Dual-Microphone Generalized Sidelobe Canceller Using General Transfer Function Jinsoo PARK, Wooil KIM,
More informationCOMPARISON OF TWO BINAURAL BEAMFORMING APPROACHES FOR HEARING AIDS
COMPARISON OF TWO BINAURAL BEAMFORMING APPROACHES FOR HEARING AIDS Elior Hadad, Daniel Marquardt, Wenqiang Pu 3, Sharon Gannot, Simon Doclo, Zhi-Quan Luo, Ivo Merks 5 and Tao Zhang 5 Faculty of Engineering,
More informationA BINAURAL HEARING AID SPEECH ENHANCEMENT METHOD MAINTAINING SPATIAL AWARENESS FOR THE USER
A BINAURAL EARING AID SPEEC ENANCEMENT METOD MAINTAINING SPATIAL AWARENESS FOR TE USER Joachim Thiemann, Menno Müller and Steven van de Par Carl-von-Ossietzky University Oldenburg, Cluster of Excellence
More informationChapter 4 SPEECH ENHANCEMENT
44 Chapter 4 SPEECH ENHANCEMENT 4.1 INTRODUCTION: Enhancement is defined as improvement in the value or Quality of something. Speech enhancement is defined as the improvement in intelligibility and/or
More informationJOINT DOA AND FUNDAMENTAL FREQUENCY ESTIMATION METHODS BASED ON 2-D FILTERING
18th European Signal Processing Conference (EUSIPCO-20) Aalborg, Denmark, August 23-27, 20 JOINT DOA AND FUNDAMENTA FREQUENCY ESTIMATION METHODS BASED ON 2-D FITERING Jesper Rindom Jensen, Mads Græsbøll
More informationCalibration of Microphone Arrays for Improved Speech Recognition
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Calibration of Microphone Arrays for Improved Speech Recognition Michael L. Seltzer, Bhiksha Raj TR-2001-43 December 2001 Abstract We present
More informationIN DISTANT speech communication scenarios, where the
IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 26, NO. 6, JUNE 2018 1119 Linear Prediction-Based Online Dereverberation and Noise Reduction Using Alternating Kalman Filters Sebastian
More informationSpeech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter
Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter 1 Gupteswar Sahu, 2 D. Arun Kumar, 3 M. Bala Krishna and 4 Jami Venkata Suman Assistant Professor, Department of ECE,
More informationSpeech and Audio Processing Recognition and Audio Effects Part 3: Beamforming
Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming Gerhard Schmidt Christian-Albrechts-Universität zu Kiel Faculty of Engineering Electrical Engineering and Information Engineering
More informationBlind Dereverberation of Single-Channel Speech Signals Using an ICA-Based Generative Model
Blind Dereverberation of Single-Channel Speech Signals Using an ICA-Based Generative Model Jong-Hwan Lee 1, Sang-Hoon Oh 2, and Soo-Young Lee 3 1 Brain Science Research Center and Department of Electrial
More informationNOISE ESTIMATION IN A SINGLE CHANNEL
SPEECH ENHANCEMENT FOR CROSS-TALK INTERFERENCE by Levent M. Arslan and John H.L. Hansen Robust Speech Processing Laboratory Department of Electrical Engineering Box 99 Duke University Durham, North Carolina
More informationarxiv: v1 [cs.sd] 17 Dec 2018
CIRCULAR STATISTICS-BASED LOW COMPLEXITY DOA ESTIMATION FOR HEARING AID APPLICATION L. D. Mosgaard, D. Pelegrin-Garcia, T. B. Elmedyb, M. J. Pihl, P. Mowlaee Widex A/S, Nymøllevej 6, DK-3540 Lynge, Denmark
More informationSPEECH signals are inherently sparse in the time and frequency
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 7, SEPTEMBER 2011 2159 An Integrated Solution for Online Multichannel Noise Tracking Reduction Mehrez Souden, Member, IEEE, Jingdong
More informationRecent advances in noise reduction and dereverberation algorithms for binaural hearing aids
Recent advances in noise reduction and dereverberation algorithms for binaural hearing aids Prof. Dr. Simon Doclo University of Oldenburg, Dept. of Medical Physics and Acoustics and Cluster of Excellence
More informationDual-Microphone Speech Dereverberation using a Reference Signal Habets, E.A.P.; Gannot, S.
DualMicrophone Speech Dereverberation using a Reference Signal Habets, E.A.P.; Gannot, S. Published in: Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP
More informationBroadband Microphone Arrays for Speech Acquisition
Broadband Microphone Arrays for Speech Acquisition Darren B. Ward Acoustics and Speech Research Dept. Bell Labs, Lucent Technologies Murray Hill, NJ 07974, USA Robert C. Williamson Dept. of Engineering,
More informationPATH UNCERTAINTY ROBUST BEAMFORMING. Richard Stanton and Mike Brookes. Imperial College London {rs408,
PATH UNCERTAINTY ROBUST BEAMFORMING Richard Stanton and Mike Brookes Imperial College London {rs8, mike.brookes}@imperial.ac.uk ABSTRACT Conventional beamformer design assumes that the phase differences
More informationHigh-speed Noise Cancellation with Microphone Array
Noise Cancellation a Posteriori Probability, Maximum Criteria Independent Component Analysis High-speed Noise Cancellation with Microphone Array We propose the use of a microphone array based on independent
More informationLocal Relative Transfer Function for Sound Source Localization
Local Relative Transfer Function for Sound Source Localization Xiaofei Li 1, Radu Horaud 1, Laurent Girin 1,2, Sharon Gannot 3 1 INRIA Grenoble Rhône-Alpes. {firstname.lastname@inria.fr} 2 GIPSA-Lab &
More informationUplink and Downlink Beamforming for Fading Channels. Mats Bengtsson and Björn Ottersten
Uplink and Downlink Beamforming for Fading Channels Mats Bengtsson and Björn Ottersten 999-02-7 In Proceedings of 2nd IEEE Signal Processing Workshop on Signal Processing Advances in Wireless Communications,
More informationIn air acoustic vector sensors for capturing and processing of speech signals
University of Wollongong Research Online University of Wollongong Thesis Collection University of Wollongong Thesis Collections 2011 In air acoustic vector sensors for capturing and processing of speech
More informationIntroduction to distributed speech enhancement algorithms for ad hoc microphone arrays and wireless acoustic sensor networks
Introduction to distributed speech enhancement algorithms for ad hoc microphone arrays and wireless acoustic sensor networks Part I: Array Processing in Acoustic Environments Sharon Gannot 1 and Alexander
More informationDirection of Arrival Algorithms for Mobile User Detection
IJSRD ational Conference on Advances in Computing and Communications October 2016 Direction of Arrival Algorithms for Mobile User Detection Veerendra 1 Md. Bakhar 2 Kishan Singh 3 1,2,3 Department of lectronics
More informationFrequency Domain Analysis for Noise Suppression Using Spectral Processing Methods for Degraded Speech Signal in Speech Enhancement
Frequency Domain Analysis for Noise Suppression Using Spectral Processing Methods for Degraded Speech Signal in Speech Enhancement 1 Zeeshan Hashmi Khateeb, 2 Gopalaiah 1,2 Department of Instrumentation
More informationMichael Brandstein Darren Ward (Eds.) Microphone Arrays. Signal Processing Techniques and Applications. With 149 Figures. Springer
Michael Brandstein Darren Ward (Eds.) Microphone Arrays Signal Processing Techniques and Applications With 149 Figures Springer Contents Part I. Speech Enhancement 1 Constant Directivity Beamforming Darren
More informationA COHERENCE-BASED ALGORITHM FOR NOISE REDUCTION IN DUAL-MICROPHONE APPLICATIONS
18th European Signal Processing Conference (EUSIPCO-21) Aalborg, Denmark, August 23-27, 21 A COHERENCE-BASED ALGORITHM FOR NOISE REDUCTION IN DUAL-MICROPHONE APPLICATIONS Nima Yousefian, Kostas Kokkinakis
More informationSPEECH ENHANCEMENT BASED ON A LOG-SPECTRAL AMPLITUDE ESTIMATOR AND A POSTFILTER DERIVED FROM CLEAN SPEECH CODEBOOK
18th European Signal Processing Conference (EUSIPCO-2010) Aalborg, Denmar, August 23-27, 2010 SPEECH ENHANCEMENT BASED ON A LOG-SPECTRAL AMPLITUDE ESTIMATOR AND A POSTFILTER DERIVED FROM CLEAN SPEECH CODEBOOK
More informationAiro Interantional Research Journal September, 2013 Volume II, ISSN:
Airo Interantional Research Journal September, 2013 Volume II, ISSN: 2320-3714 Name of author- Navin Kumar Research scholar Department of Electronics BR Ambedkar Bihar University Muzaffarpur ABSTRACT Direction
More informationBREAKING DOWN THE COCKTAIL PARTY: CAPTURING AND ISOLATING SOURCES IN A SOUNDSCAPE
BREAKING DOWN THE COCKTAIL PARTY: CAPTURING AND ISOLATING SOURCES IN A SOUNDSCAPE Anastasios Alexandridis, Anthony Griffin, and Athanasios Mouchtaris FORTH-ICS, Heraklion, Crete, Greece, GR-70013 University
More informationEigenvalues and Eigenvectors in Array Antennas. Optimization of Array Antennas for High Performance. Self-introduction
Short Course @ISAP2010 in MACAO Eigenvalues and Eigenvectors in Array Antennas Optimization of Array Antennas for High Performance Nobuyoshi Kikuma Nagoya Institute of Technology, Japan 1 Self-introduction
More informationSTAP approach for DOA estimation using microphone arrays
STAP approach for DOA estimation using microphone arrays Vera Behar a, Christo Kabakchiev b, Vladimir Kyovtorov c a Institute for Parallel Processing (IPP) Bulgarian Academy of Sciences (BAS), behar@bas.bg;
More informationTitle. Author(s)Sugiyama, Akihiko; Kato, Masanori; Serizawa, Masahir. Issue Date Doc URL. Type. Note. File Information
Title A Low-Distortion Noise Canceller with an SNR-Modifie Author(s)Sugiyama, Akihiko; Kato, Masanori; Serizawa, Masahir Proceedings : APSIPA ASC 9 : Asia-Pacific Signal Citationand Conference: -5 Issue
More informationSound Source Localization using HRTF database
ICCAS June -, KINTEX, Gyeonggi-Do, Korea Sound Source Localization using HRTF database Sungmok Hwang*, Youngjin Park and Younsik Park * Center for Noise and Vibration Control, Dept. of Mech. Eng., KAIST,
More informationSubspace Noise Estimation and Gamma Distribution Based Microphone Array Post-filter Design
Chinese Journal of Electronics Vol.0, No., Apr. 011 Subspace Noise Estimation and Gamma Distribution Based Microphone Array Post-filter Design CHENG Ning 1,,LIUWenju 3 and WANG Lan 1, (1.Shenzhen Institutes
More informationMMSE STSA Based Techniques for Single channel Speech Enhancement Application Simit Shah 1, Roma Patel 2
MMSE STSA Based Techniques for Single channel Speech Enhancement Application Simit Shah 1, Roma Patel 2 1 Electronics and Communication Department, Parul institute of engineering and technology, Vadodara,
More informationSpeech Enhancement using Wiener filtering
Speech Enhancement using Wiener filtering S. Chirtmay and M. Tahernezhadi Department of Electrical Engineering Northern Illinois University DeKalb, IL 60115 ABSTRACT The problem of reducing the disturbing
More informationSpeech enhancement with ad-hoc microphone array using single source activity
Speech enhancement with ad-hoc microphone array using single source activity Ryutaro Sakanashi, Nobutaka Ono, Shigeki Miyabe, Takeshi Yamada and Shoji Makino Graduate School of Systems and Information
More informationSpringer Topics in Signal Processing
Springer Topics in Signal Processing Volume 3 Series Editors J. Benesty, Montreal, Québec, Canada W. Kellermann, Erlangen, Germany Springer Topics in Signal Processing Edited by J. Benesty and W. Kellermann
More informationStudy Of Sound Source Localization Using Music Method In Real Acoustic Environment
International Journal of Electronics Engineering Research. ISSN 975-645 Volume 9, Number 4 (27) pp. 545-556 Research India Publications http://www.ripublication.com Study Of Sound Source Localization Using
More informationSpeech Enhancement using Multiple Transducers
Speech Enhancement using Multiple Transducers Craig Anderson A Thesis submitted to the Victoria University of Wellington in fulfilment of the requirements for the degree of Master of Engineering Victoria
More informationIndoor Localization based on Multipath Fingerprinting. Presented by: Evgeny Kupershtein Instructed by: Assoc. Prof. Israel Cohen and Dr.
Indoor Localization based on Multipath Fingerprinting Presented by: Evgeny Kupershtein Instructed by: Assoc. Prof. Israel Cohen and Dr. Mati Wax Research Background This research is based on the work that
More informationENERGY-VS-PERFORMANCE TRADE-OFFS IN SPEECH ENHANCEMENT IN WIRELESS ACOUSTIC SENSOR NETWORKS
ENERGY-VS-PERFORMANCE TRADE-OFFS IN SPEECH ENHANCEMENT IN WIRELESS ACOUSTIC SENSOR NETWORKS Fernando de la Hucha Arce 1, Fernando Rosas, Marc Moonen 1, Marian Verhelst, Alexander Bertrand 1 KU Leuven,
More informationAdaptive Systems Homework Assignment 3
Signal Processing and Speech Communication Lab Graz University of Technology Adaptive Systems Homework Assignment 3 The analytical part of your homework (your calculation sheets) as well as the MATLAB
More informationSpeech Signal Enhancement Techniques
Speech Signal Enhancement Techniques Chouki Zegar 1, Abdelhakim Dahimene 2 1,2 Institute of Electrical and Electronic Engineering, University of Boumerdes, Algeria inelectr@yahoo.fr, dahimenehakim@yahoo.fr
More informationJoint DOA and Array Manifold Estimation for a MIMO Array Using Two Calibrated Antennas
1 Joint DOA and Array Manifold Estimation for a MIMO Array Using Two Calibrated Antennas Wei Zhang #, Wei Liu, Siliang Wu #, and Ju Wang # # Department of Information and Electronics Beijing Institute
More informationDifferent Approaches of Spectral Subtraction Method for Speech Enhancement
ISSN 2249 5460 Available online at www.internationalejournals.com International ejournals International Journal of Mathematical Sciences, Technology and Humanities 95 (2013 1056 1062 Different Approaches
More informationADAPTIVE ANTENNAS. TYPES OF BEAMFORMING
ADAPTIVE ANTENNAS TYPES OF BEAMFORMING 1 1- Outlines This chapter will introduce : Essential terminologies for beamforming; BF Demonstrating the function of the complex weights and how the phase and amplitude
More informationNoise Reduction: An Instructional Example
Noise Reduction: An Instructional Example VOCAL Technologies LTD July 1st, 2012 Abstract A discussion on general structure of noise reduction algorithms along with an illustrative example are contained
More informationSPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN. Yu Wang and Mike Brookes
SPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN Yu Wang and Mike Brookes Department of Electrical and Electronic Engineering, Exhibition Road, Imperial College London,
More informationTowards an intelligent binaural spee enhancement system by integrating me signal extraction. Author(s)Chau, Duc Thanh; Li, Junfeng; Akagi,
JAIST Reposi https://dspace.j Title Towards an intelligent binaural spee enhancement system by integrating me signal extraction Author(s)Chau, Duc Thanh; Li, Junfeng; Akagi, Citation 2011 International
More information