Recent Advances in Acoustic Signal Extraction and Dereverberation

Similar documents
Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor. Presented by Amir Kiperwas

Informed Spatial Filtering for Sound Extraction Using Distributed Microphone Arrays

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 24, NO. 7, JULY

Flexible and efficient spatial sound acquisition and subsequent. Parametric Spatial Sound Processing

Dual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation

Multiple Sound Sources Localization Using Energetic Analysis Method

Recent advances in noise reduction and dereverberation algorithms for binaural hearing aids

Speech Enhancement Using Microphone Arrays

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment

Enhancement of Speech Signal Based on Improved Minima Controlled Recursive Averaging and Independent Component Analysis

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm

arxiv: v1 [cs.sd] 4 Dec 2018

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.

Linear and Parametric Microphone Array Processing

The Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals

NOISE POWER SPECTRAL DENSITY MATRIX ESTIMATION BASED ON MODIFIED IMCRA. Qipeng Gong, Benoit Champagne and Peter Kabal

BREAKING DOWN THE COCKTAIL PARTY: CAPTURING AND ISOLATING SOURCES IN A SOUNDSCAPE

Joint dereverberation and residual echo suppression of speech signals in noisy environments Habets, E.A.P.; Gannot, S.; Cohen, I.; Sommen, P.C.W.

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming

/$ IEEE

Performance Analysis of Acoustic Echo Cancellation in Sound Processing

Microphone Array Power Ratio for Speech Quality Assessment in Noisy Reverberant Environments 1

Joint recognition and direction-of-arrival estimation of simultaneous meetingroom acoustic events

A BINAURAL HEARING AID SPEECH ENHANCEMENT METHOD MAINTAINING SPATIAL AWARENESS FOR THE USER

All-Neural Multi-Channel Speech Enhancement

Antennas and Propagation. Chapter 5c: Array Signal Processing and Parametric Estimation Techniques

arxiv: v1 [cs.sd] 17 Dec 2018

Towards an intelligent binaural spee enhancement system by integrating me signal extraction. Author(s)Chau, Duc Thanh; Li, Junfeng; Akagi,

Different Approaches of Spectral Subtraction Method for Speech Enhancement

546 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 4, MAY /$ IEEE

The psychoacoustics of reverberation

DISTANT or hands-free audio acquisition is required in

MMSE STSA Based Techniques for Single channel Speech Enhancement Application Simit Shah 1, Roma Patel 2

MULTICHANNEL AUDIO DATABASE IN VARIOUS ACOUSTIC ENVIRONMENTS

High-speed Noise Cancellation with Microphone Array

REAL TIME DIGITAL SIGNAL PROCESSING

IN REVERBERANT and noisy environments, multi-channel

ROOM AND CONCERT HALL ACOUSTICS MEASUREMENTS USING ARRAYS OF CAMERAS AND MICROPHONES

260 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 2, FEBRUARY /$ IEEE

In air acoustic vector sensors for capturing and processing of speech signals

Chapter 4 SPEECH ENHANCEMENT

Performance Evaluation of Nonlinear Speech Enhancement Based on Virtual Increase of Channels in Reverberant Environments

A BROADBAND BEAMFORMER USING CONTROLLABLE CONSTRAINTS AND MINIMUM VARIANCE

Improving reverberant speech separation with binaural cues using temporal context and convolutional neural networks

TARGET SPEECH EXTRACTION IN COCKTAIL PARTY BY COMBINING BEAMFORMING AND BLIND SOURCE SEPARATION

Advances in Direction-of-Arrival Estimation

ADAPTIVE ANTENNAS. TYPES OF BEAMFORMING

Calibration of Microphone Arrays for Improved Speech Recognition

Microphone Array Design and Beamforming

A HYPOTHESIS TESTING APPROACH FOR REAL-TIME MULTICHANNEL SPEECH SEPARATION USING TIME-FREQUENCY MASKS. Ryan M. Corey and Andrew C.

Smart antenna for doa using music and esprit

Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks

Speech Enhancement for Nonstationary Noise Environments

Frequency Domain Analysis for Noise Suppression Using Spectral Processing Methods for Degraded Speech Signal in Speech Enhancement

RECENTLY, there has been an increasing interest in noisy

Indoor Localization based on Multipath Fingerprinting. Presented by: Evgeny Kupershtein Instructed by: Assoc. Prof. Israel Cohen and Dr.

Simultaneous Recognition of Speech Commands by a Robot using a Small Microphone Array

Adaptive Systems Homework Assignment 3

Performance Analysis of MUSIC and MVDR DOA Estimation Algorithm

Microphone Array Feedback Suppression. for Indoor Room Acoustics

SPECTRAL COMBINING FOR MICROPHONE DIVERSITY SYSTEMS

Effects of Reverberation on Pitch, Onset/Offset, and Binaural Cues

Clustered Multi-channel Dereverberation for Ad-hoc Microphone Arrays

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 5, MAY

Online Version Only. Book made by this file is ILLEGAL. 2. Mathematical Description

Improving Meetings with Microphone Array Algorithms. Ivan Tashev Microsoft Research

MULTICHANNEL systems are often used for

ESTIMATION OF TIME-VARYING ROOM IMPULSE RESPONSES OF MULTIPLE SOUND SOURCES FROM OBSERVED MIXTURE AND ISOLATED SOURCE SIGNALS

Estimation of Reverberation Time from Binaural Signals Without Using Controlled Excitation

Michael Brandstein Darren Ward (Eds.) Microphone Arrays. Signal Processing Techniques and Applications. With 149 Figures. Springer

Reverberant Sound Localization with a Robot Head Based on Direct-Path Relative Transfer Function

GROUP SPARSITY FOR MIMO SPEECH DEREVERBERATION. and the Cluster of Excellence Hearing4All, Oldenburg, Germany.

Robust Low-Resource Sound Localization in Correlated Noise

BEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR

ROBUST echo cancellation requires a method for adjusting

Springer Topics in Signal Processing

Direction-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method

A generalized framework for binaural spectral subtraction dereverberation

Binaural Beamforming with Spatial Cues Preservation

Proceedings of Meetings on Acoustics

AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION

AN ADAPTIVE MICROPHONE ARRAY FOR OPTIMUM BEAMFORMING AND NOISE REDUCTION

HUMAN speech is frequently encountered in several

Automotive three-microphone voice activity detector and noise-canceller

Dual-Microphone Speech Dereverberation in a Noisy Environment

Adaptive Filters Wiener Filter

Advanced Signal Processing and Digital Noise Reduction

Monaural and Binaural Speech Separation

Two-channel Separation of Speech Using Direction-of-arrival Estimation And Sinusoids Plus Transients Modeling

PRIMARY-AMBIENT SOURCE SEPARATION FOR UPMIXING TO SURROUND SOUND SYSTEMS

A MULTI-CHANNEL POSTFILTER BASED ON THE DIFFUSE NOISE SOUND FIELD. Lukas Pfeifenberger 1 and Franz Pernkopf 1

Spectral Methods for Single and Multi Channel Speech Enhancement in Multi Source Environment

Noise Reduction for L-3 Nautronix Receivers

On Single-Channel Speech Enhancement and On Non-Linear Modulation-Domain Kalman Filtering

Lecture 14: Source Separation

Effective post-processing for single-channel frequency-domain speech enhancement Weifeng Li a

Mel Spectrum Analysis of Speech Recognition using Single Microphone

A classification-based cocktail-party processor

Speech Signal Enhancement Techniques

Transcription:

Recent Advances in Acoustic Signal Extraction and Dereverberation Emanuël Habets Erlangen Colloquium 2016

Scenario Spatial Filtering Estimated Desired Signal Undesired sound components: Sensor noise Competing speakers Ambient sounds (e.g., generated by an air conditioner, fan, or babble) Reverberation (due to wall reflections, etc.) 2

Applications Hands- free Communication Human- Machine Interfaces Hearing Aids Music Recording and Post- Production AudioLabs, 2016 Emanuël Habets Acoustic Signal Extraction and Dereverberation 3

Outline Acoustic Signal Extraction Dereverberation Reverberation Cancellation Reverberation Suppression Conclusions and Future Challenges 4

Outline Acoustic Signal Extraction Dereverberation Reverberation Cancellation Reverberation Suppression Conclusions and Future Challenges 5

Acoustic Signal Extraction Goal Extract the desired signal while reducing undesired signals from one or more microphone signals Solutions Single-channel filters Data-independent beamformers Data-dependent beamformers Challenges Defining the desired signal Estimating the spatio-temporal statistics in non-stationary scenarios 6

Acoustic Signal Extraction We developed two approaches that use nearly instantaneous information about the acoustic scene to overcome the challenges of estimating the spatio-temporal statistics We refer to these as informed spatial filtering approaches The main difference lies in the way the spatial information is used The direct approach uses the spatial information to control the filters The indirect approach uses the spatial information to distinguish between desired and undesired sounds 7

Direct Informed Spatial Filtering Parametric Sound Field Model The total sound field is formed as a superposition of the direct sound field and diffuse sound field microphone signals direct sound diffuse sound total sound field In practice, the DOA of the direct sound can vary quickly, for instance, when multiple talkers are active at the same time 8

Direct Informed Spatial Filtering Parametric Sound Field Model In the TF domain the microphone signals can be expressed as: relative TFs diffuse sound l-th plane wave stationary noise (e.g., fan, sensor-noise) The diffuse sound power varies quickly across time. The PSD matrix of the diffuse sound component can be expressed as coherence matrix: time-invariant and known diffuse sound power The statistics of the background noise vary slowly across time and can be estimated from the microphone signals 9

Direct Informed Spatial Filtering Desired Signal Our objective is to capture L plane waves with desired gains while suppressing diffuse sound and noise (Thiergart et al., 2014) The desired signal is given by desired spatial response for the l-th plane wave 0 5 10 15 G1 G2 Desired Spatial Responses 20 90 45 0 45 90 DOA 10

Direct Informed Spatial Filtering Estimation of the Desired Signal The desired signal is estimated using a spatial filter The informed LCMV filter is for example given by Residual Noise plus Reverberation The required narrowband DOAs of the plane waves can be estimated using ESPRIT or root-music An estimator for the DNR was proposed in (Thiergart et al., 2014) 11

Direct Informed Spatial Filtering Estimation of the Desired Signal The desired signal is estimated using a spatial filter The informed LCMV filter is for example given by Diffuse-to-Noise Ratio The required narrowband DOAs of the plane waves can be estimated using ESPRIT or root-music An estimator for the DNR was proposed in (Thiergart et al., 2014) 12

Direct Informed Spatial Filtering Proposed System DNR Estimation DOA Estimation Filter Weights Desired Response frequency [khz] 5 4 3 2 1 0 0.5 1 1.5 2 2.5 3 time [s] 90 45 0 45 90 frequency [khz] 5 4 3 2 1 0 0.5 1 1.5 2 2.5 3 time [s] 0 5 10 15 20 13

Direct Informed Spatial Filtering Example with Face Tracking The demo can be found at https://www.audiolabs-erlangen.de/resources/2015-mcse 14

Direct Informed Spatial Filtering Conclusions and Current Work A flexible spatial filtering approach that can be used to realize different audio applications independent of the microphone setup The approach offer a high robustness in quickly changing acoustic scenarios The achievable spatial selectively depends on the DOA accuracy Current Work Developing even more robust parameter estimators Incorporating DOA uncertainties Applying this approach to binaural hearing aids 15

Indirect Informed Spatial Filtering An alternative approach was developed that aims in particular at acoustic signal extraction In this case instantaneous information about the acoustic scene is used to classify each TF instance as desired or undesired Microphone Signals Spatial Filtering Processed Signal Acoustic Scene Analysis Classification Spatio- Temporal Statistics Estimation 16

Indirect Informed Spatial Filtering Signal Model In the TF domain the m-th microphone signal can be expressed as Input vector: Desired PSD matrix: Undesired PSD matrix: 17

Indirect Informed Spatial Filtering Power Spectral Density Estimation Using the estimated PSD matrices optimal filters can be computed Minimum variance distortionless response filter Multi-channel Wiener filter Estimate the PSD matrix (for each frequency bin k) Each TF instance can be classified as Desired signal present Desired signal absent The smoothing constants depend on the classification 18

Indirect Informed Spatial Filtering Minimum Bayes-risk Detector We propose to classify each TF instance using spatial features and a minimum Bayes-risk decision rule Probabilities of the hypotheses given the estimated features Bayes costs control the tradeoff between speech distortion and interference reduction How to obtain the posterior probabilities of the hypotheses? 19

Indirect Informed Spatial Filtering DOA-based Minimum Bayes-risk Detector Let us assume a single microphone array and known target direction We propose to classify each TF instance using narrowband directionof-arrivals (DOA) and signal-to-diffuse ratio (SDR) estimates Mixture model for the estimated DOAs Posterior probability of the hypothesis uniformly distributed Source: (Taseska and Habets, 2015) 20

Indirect Informed Spatial Filtering DOA-based Minimum Bayes-risk Detector f(ˆ H d ;0, apple) apple =1 apple =3 apple = 10 modes at the target DOA pi pi/2 0 pi/2 pi pi pi/2 0 pi/2 pi Von Mises distribution 2 Desired source 0 varies based on direct-to-diffuse ratio 2 approximately uniform region low values near the target DOA Notched distribution 2 f(ˆ H i ;0, apple) 0 apple =1 apple =3 apple = 10 2 Mode corresponds to the target DOA The concentration parameter reflects the DOA estimator uncertainty 21

Indirect Informed Spatial Filtering DOA-based Minimum Bayes-risk Detector Estimate DOA Target DOA Bayes Costs Estimate SDR Compute Concentration Parameter Estimate SPP Compute Likelihoods Compute Mixture Coefficients Minimum Bayes-Risk Detector Decision The mixture coefficients (i.e., prior probabilities) are computed using the speech presence probability 22

Indirect Informed Spatial Filtering Example Using DOA Model-based Detector Setup Sampling frequency 16 khz STFT frames 64 ms, 50% overlap Circular array (3 DPAs, 1.5 cm radius) Reverberation time 0.18 s Sensor and diffuse noise Signal-to-interference ratio approx. 3 db Acoustic Signal Extraction (Taseska and Habets, 2015) Minimum variance distortionless response (MVDR) beamformer PSD matrices are estimated using three different detectors 23

Indirect Informed Spatial Filtering Example Using DOA Model-based Detector Reference microphone Source1 Source 2 Ideal detector Signal model-based detector (Jarrett et al., 2014) DOA model based detector (Taseska and Habets, 2015) Some audio demos can be found at https://www.audiolabs-erlangen.de/resources/2015-icassp-doadet 24

Indirect Informed Spatial Filtering Spotforming (Taseska and Habets, 2013) Beamforming Spotforming Interferer Interferers Interferers Source of interest Noise source Sources of interest Noise source 25

Indirect Informed Spatial Filtering Position-based Minimum Bayes-risk Detector Using distributed arrays narrowband position estimates can be obtained These positions are used as a spatial feature for the classification conditional spot probability desired speaker speech presence probability speech undesired speaker non-speech 26

Indirect Informed Spatial Filtering Position-based Minimum Bayes-risk Detector The conditional spot probability is given by Gaussian likelihood model Uniform prior 27

Indirect Informed Spatial Filtering Example Using Position Model-based Detector Reverberation time approx. 0.18 s Three circular arrays, three DPA microphones per array, 3 cm diameter 16 khz sampling rate, STFT frame 64 ms with 50 % overlap Spotformer using an MVDR filter Comparison using a oracle fixed spotformer where the PSD matrices were optimal during the first 10 seconds Some audio demos can be found at https://www.audiolabs-erlangen.de/resources/2015-spotformer 28

Indirect Informed Spatial Filtering Example Using Position Model-based Detector Reference Microphone 29

Indirect Informed Spatial Filtering Example Using Position Model-based Detector Oracle Fixed Spotformer 30

Indirect Informed Spatial Filtering Example Using Position Model-based Detector Proposed Data-Dependent Spotformer 31

Indirect Informed Spatial Filtering Example Using Position Model-based Detector Scenario: moving sources Reverberation time: 0.3 s Scene analysis: 9 mics Spatial filtering: 3 mics SIR Input 0 db SIR Output 12.1 db SNR Input 0 db SNR Output 6.8 db 32

Indirect Informed Spatial Filtering Conclusions and Current Work The indirect ISF approach can be used to extract acoustic signals arriving from a specific direction or location Provides low speech distortion and high interference reduction Current Work Developing even more robust detectors Performing distributed signal processing 33

Outline Acoustic Signal Extraction Dereverberation* Reverberation Cancellation Reverberation Suppression Conclusions and Future Challenges * In collaboration with Sharon Gannot, Boaz Schwartz, and Ofer Schwartz from Bar-Ilan University, Israel 34

Dereverberation Publications 120 Start of my PhD End of my PhD 100 80 60 40 20 0 1935 1945 1955 1965 1975 1985 1995 2005 2015 Without ASR With ASR Source: Scopus 35

Dereverberation Approaches Three fundamentally different approaches 1. Model the acoustic system, estimate the model parameters by treading the source signal as a nuisance, and then estimate the source signal 2. Model the reverberation as an additive process, and then estimate the source signal 3. Directly estimate the source signal from the microphone signals by treading the acoustic system as unknown Source Signal Microphone Signals Enhanced Signal Acoustic System Dereverberation 36

Dereverberation Approaches Three fundamentally different approaches 1. Model the acoustic system, estimate the model parameters by treading the source signal Reverberation as a nuisance, and then Cancellation estimate the source signal 2. Model the reverberation as an additive process, and then estimate the source signal 3. Directly estimate the source signal from the microphone signals by treading the acoustic system as unknown Source Signal Microphone Signals Enhanced Signal Acoustic System Dereverberation 37

Dereverberation Approaches Three fundamentally different approaches 1. Model the acoustic system, estimate the model parameters by treading the source signal as a nuisance, and then estimate the source signal 2. Model the reverberation as an additive process, and then estimate the source signal 3. Directly estimate the source signal from the microphone signals by treading the acoustic system as unknown Reverberation Source Signal + Microphone Signal Dereverberation Enhanced Signal 38

Dereverberation Approaches Three fundamentally different approaches 1. Model the acoustic system, estimate the model parameters by treading the source signal as a nuisance, and then estimate the source signal 2. Model the reverberation as an additive process, and then estimate the source signal Reverberation Suppression 3. Directly estimate the source signal from the microphone signals by treading the acoustic system as unknown Reverberation Source Signal + Microphone Signal Dereverberation Enhanced Signal 39

Dereverberation Approaches Three fundamentally different approaches 1. Model the acoustic system, estimate the model parameters by treading the source signal as a nuisance, and then estimate the source signal 2. Model the reverberation as an additive process, and then estimate the source signal 3. Directly estimate the source signal from the microphone signals by treading the acoustic system as unknown Source Signal Microphone Signals? Dereverberation Enhanced Signal 40

Dereverberation Approaches Three fundamentally different approaches 1. Model the acoustic system, estimate the model parameters by treading the source signal as a nuisance, and then estimate the source signal 2. Model the reverberation as an additive process, and then estimate the source signal 3. Directly estimate the source signal from the microphone signals by treading the acoustic Direct system as Estimation unknown Source Signal Microphone Signals? Dereverberation Enhanced Signal 41

Reverberation Cancellation Models Acoustic models Finite Impulse Response Source Signal Microphone Signal Infinite Impulse Response Signal models h(0,n) Moving average process Autoregressive process The models can be described in the time-domain or time-frequency domain 42

Reverberation Cancellation Moving Average Process (Time-Domain) The desired signal is a delayed or filtered version of the source signal To obtain an estimate of the desired signal: 1. Blindly identify the model parameters of the acoustic system 2. Estimate the desired signal by applying a multichannel equalizer Source Signal Acoustic System Microphone Signals Multichannel Equalization Enhanced Signal Blind System Identification 43

Reverberation Cancellation Moving Average Process (TF-Domain) In (B. Schwartz et al., 2015) the microphone signals were modeled in the TF domain as a moving average process In the context of binaural hearing aids (B. Schwartz et al., 2015): 44

Reverberation Cancellation Moving Average Process (TF-Domain) In (B. Schwartz et al., 2015) the microphone signals were modeled in the TF domain as a moving average process In the context of binaural hearing aids (B. Schwartz et al., 2015): 45

Reverberation Cancellation Moving Average Process (TF-Domain) In (B. Schwartz et al., 2015) the microphone signals were modeled in the TF domain as a moving average process In the context of binaural hearing aids (B. Schwartz et al., 2015): Early Speech at the Reference Microphone Relative CTFs 46

Reverberation Cancellation Moving Average Process (TF-Domain) A recursive expectationmaximization scheme is used to estimate online the acoustic system, speech, and noise parameters In the E-Step, a Kalman filter is used to estimate the desired speech signal (and the error covariance matrix) Source: (B. Schwartz et al., 2015) 47

Reverberation Cancellation Binaural Hearing Aids Source: (B. Schwartz et al., 2015) 48

Reverberation Cancellation Binaural Hearing Aids WSNR Improvement (db) 4 3.5 3 2.5 2 1.5 1 α=0.1 α=0.35 α=0.7 α= [-10,-6] [-6,-4] [-4,-2.3] [-2.3,0.5] [0.5,4] DRR range (db) Source: (B. Schwartz et al., 2015) 49

Reverberation Cancellation Binaural Hearing Aids ITD and ILD distribution for the different reverberation levels and window functions These plots relate to Position 3 Source: (B. Schwartz et al., 2015) 50

Reverberation Suppression It is assumed that 1. reverberation is an additive process 2. the desired signal and the reverberant signal are uncorrelated 3. the reverberant signal can be modeled as r(n, k) N C (0, R(n, r(n, k)) (k)) 51

Reverberation Suppression Single-Channel Spectral Enhancement Single-channel spectral enhancement techniques commonly require an estimate of the clean speech PSD and the interference PSD Statistical models for the acoustic channel can be used to derive estimators for the reverberation PSD Microphone Signal TF Analysis Estimate Early Speech Signal Synthesis Enhanced Signal Estimate Reverberation PSD prior information (RT, DRR) 52

Reverberation Suppression Single-Channel Spectral Enhancement Selected approaches to estimate the reverberation PSD: Lebart et al. (2001) used Moorer s model and a frequency independent reverberation time (RT) Habets and Sommen (2003) used Polack s model and a frequency dependent RT Habets et al. (2007/2009) proposed a generalized statistical model that depends on the direct-to-reverberation ratio (DRR) and RT Erkelens et al. (2010) proposed a correlation-based PSD estimator This let to new challenges such as blindly estimating the DRR and RT which were also part of the recent ACE 2015 Challenge 53

Reverberation Suppression Data-Dependent Spatial Filtering Fully exploit the spatial diversity of the desired and undesired signals Estimate the propagation vector and reverberation PSD matrix Microphone Signals TF Analyses Estimate Desired Signal Synthesis Enhanced Signal Estimate Propagation Vector and Rev. PSD Matrix 54

Reverberation Suppression Data-Dependent Spatial Filtering In (O. Schwartz et al., 2016) we used a multi-channel MMSE filter The relative early transfer functions, as well as the level and spatial coherence matrix of the reverberation were iteratively estimated using an Expectation-Maximization scheme Example (Distance = 2 m, RT 60 =0.61 s, Noiseless) Input Singe-channel Dereverberation (Habets, 2007) Four-channel Dereverberation (2016) Female Male Some audio demos can be found at http://www.eng.biu.ac.il/gannot/speech-enhancement/wiener-em 55

Outline Acoustic Signal Extraction Dereverberation Reverberation Cancellation Reverberation Suppression Conclusions and Future Challenges 56

Conclusions and Future Challenges Significant advances have been made in the areas of acoustic signal extraction and dereverberation Using newly developed acoustic signal processing techniques we are starting to see a true benefit of multi-microphone processing Future Challenges Lower signal-to-noise ratios and higher reverberation times Incorporating perceptual models and knowledge Automatic adaptation of the desired spatial response 57

Special thanks to Oliver Thiergart (AudioLabs) Maja Taseska (AudioLabs) Sharon Gannot (Bar-Ilan University, Israel) Boaz Schwartz (Bar-Ilan University, Israel) Ofer Schwartz (Bar-Ilan University, Israel) Thank you for your attention. 58

References O. Thiergart, M. Taseska and E.A.P. Habets, An informed parametric spatial filter based on instantaneous direction-of-arrival estimates, IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 22, Issue 12, pp. 2182-2196, Dec. 2014. K. Kowalczyk, O. Thiergart, M. Taseska, G. Del Galdo, V. Pulkki and E.A.P. Habets, "Parametric spatial sound processing: A flexible and efficient solution to sound scene acquisition, modification and reproduction," IEEE Signal Processing Magazine, Vol. 32, Issue 2, pp. 31-42, Mar. 2015. M. Taseska and E.A.P. Habet, "Minimum Bayes risk signal detection for speech enhancement based on a narrowband DOA model," Proc. of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Brisbane, Australia, 2015. M. Taseska and E.A.P. Habets, "Spotforming using distributed microphone arrays," Best Student Paper Award, Proc. of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, USA, Oct. 20-23, 2013. B. Schwartz, S. Gannot and E.A.P. Habets, An online dereverberation algorithm for hearing aids with binaural cues preservation, Proc. of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), 2015. O. Schwartz, S. Gannot and E.A.P. Habets, "An expectation-maximization algorithm for multi-microphone speech dereverberation and noise reduction with coherence matrix estimation," IEEE/ACM Transactions on Audio, Speech, and Language Processing," to appear. 59