SEPARATION AND DEREVERBERATION PERFORMANCE OF FREQUENCY DOMAIN BLIND SOURCE SEPARATION. Ryo Mukai Shoko Araki Shoji Makino

Similar documents
REAL-TIME BLIND SOURCE SEPARATION FOR MOVING SPEAKERS USING BLOCKWISE ICA AND RESIDUAL CROSSTALK SUBTRACTION

BLIND SOURCE separation (BSS) [1] is a technique for

The Role of High Frequencies in Convolutive Blind Source Separation of Speech Signals

Nonlinear postprocessing for blind speech separation

High-speed Noise Cancellation with Microphone Array

Performance Evaluation of Nonlinear Speech Enhancement Based on Virtual Increase of Channels in Reverberant Environments

Blind Dereverberation of Single-Channel Speech Signals Using an ICA-Based Generative Model

WHITENING PROCESSING FOR BLIND SEPARATION OF SPEECH SIGNALS

Source Separation and Echo Cancellation Using Independent Component Analysis and DWT

Frequency-Domain Blind Source Separation of Many Speech Signals Using Near-Field and Far-Field Models

BLIND SOURCE SEPARATION BASED ON ACOUSTIC PRESSURE DISTRIBUTION AND NORMALIZED RELATIVE PHASE USING DODECAHEDRAL MICROPHONE ARRAY

Grouping Separated Frequency Components by Estimating Propagation Model Parameters in Frequency-Domain Blind Source Separation

SUPERVISED SIGNAL PROCESSING FOR SEPARATION AND INDEPENDENT GAIN CONTROL OF DIFFERENT PERCUSSION INSTRUMENTS USING A LIMITED NUMBER OF MICROPHONES

Real-time Adaptive Concepts in Acoustics

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 4, MAY /$ IEEE

BLIND SOURCE SEPARATION FOR CONVOLUTIVE MIXTURES USING SPATIALLY RESAMPLED OBSERVATIONS

A Novel Hybrid Approach to the Permutation Problem of Frequency Domain Blind Source Separation

THE problem of acoustic echo cancellation (AEC) was

TARGET SPEECH EXTRACTION IN COCKTAIL PARTY BY COMBINING BEAMFORMING AND BLIND SOURCE SEPARATION

Separation of Noise and Signals by Independent Component Analysis

516 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING

Separation of Multiple Speech Signals by Using Triangular Microphone Array

+ C(0)21 C(1)21 Z -1. S1(t) + - C21. E1(t) C(D)21 C(D)12 C12 C(1)12. E2(t) S2(t) (a) Original H-J Network C(0)12. (b) Extended H-J Network

Multiple Sound Sources Localization Using Energetic Analysis Method

Chapter 4 SPEECH ENHANCEMENT

A Comparison of the Convolutive Model and Real Recording for Using in Acoustic Echo Cancellation

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment

Enhancement of Speech Signal Based on Improved Minima Controlled Recursive Averaging and Independent Component Analysis

Design and Implementation on a Sub-band based Acoustic Echo Cancellation Approach

Recent Advances in Acoustic Signal Extraction and Dereverberation

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter

Improving Meetings with Microphone Array Algorithms. Ivan Tashev Microsoft Research

Audiovisual speech source separation: a regularization method based on visual voice activity detection

Electronic Research Archive of Blekinge Institute of Technology

SPECTRAL COMBINING FOR MICROPHONE DIVERSITY SYSTEMS

Robust Low-Resource Sound Localization in Correlated Noise

BLIND SEPARATION OF LINEAR CONVOLUTIVE MIXTURES USING ORTHOGONAL FILTER BANKS. Milutin Stanacevic, Marc Cohen and Gert Cauwenberghs

Effective post-processing for single-channel frequency-domain speech enhancement Weifeng Li a

Automotive three-microphone voice activity detector and noise-canceller

Unidirectional Sound Signage for Speech Frequency Range Using Multiple-Loudspeaker Reproduction System

Harmonics Enhancement for Determined Blind Sources Separation using Source s Excitation Characteristics

MULTIMODAL BLIND SOURCE SEPARATION WITH A CIRCULAR MICROPHONE ARRAY AND ROBUST BEAMFORMING

ICA for Musical Signal Separation

RASTA-PLP SPEECH ANALYSIS. Aruna Bayya. Phil Kohn y TR December 1991

Convention Paper Presented at the 138th Convention 2015 May 7 10 Warsaw, Poland

FROM BLIND SOURCE SEPARATION TO BLIND SOURCE CANCELLATION IN THE UNDERDETERMINED CASE: A NEW APPROACH BASED ON TIME-FREQUENCY ANALYSIS

Speech enhancement with ad-hoc microphone array using single source activity

Multichannel Acoustic Signal Processing for Human/Machine Interfaces -

REAL TIME DIGITAL SIGNAL PROCESSING

Dual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation

ROBUST BLIND SOURCE SEPARATION IN A REVERBERANT ROOM BASED ON BEAMFORMING WITH A LARGE-APERTURE MICROPHONE ARRAY

AMAIN cause of speech degradation in practically all listening

ESTIMATION OF TIME-VARYING ROOM IMPULSE RESPONSES OF MULTIPLE SOUND SOURCES FROM OBSERVED MIXTURE AND ISOLATED SOURCE SIGNALS

Sound Processing Technologies for Realistic Sensations in Teleworking

Three-dimensional sound field simulation using the immersive auditory display system Sound Cask for stage acoustics

WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY

Discrete Fourier Transform (DFT)

Single Channel Speaker Segregation using Sinusoidal Residual Modeling

Chapter 4 DOA Estimation Using Adaptive Array Antenna in the 2-GHz Band

Joint recognition and direction-of-arrival estimation of simultaneous meetingroom acoustic events

Adaptive Filters Application of Linear Prediction

Calibration of Microphone Arrays for Improved Speech Recognition

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm

Underdetermined Convolutive Blind Source Separation via Frequency Bin-wise Clustering and Permutation Alignment

Speech Enhancement Using Microphone Arrays

MINUET: MUSICAL INTERFERENCE UNMIXING ESTIMATION TECHNIQUE

Missing-Feature based Speech Recognition for Two Simultaneous Speech Signals Separated by ICA with a pair of Humanoid Ears

MURDOCH RESEARCH REPOSITORY

Evaluation of clipping-noise suppression of stationary-noisy speech based on spectral compensation

About Multichannel Speech Signal Extraction and Separation Techniques

Speech Enhancement Based On Noise Reduction

Image De-Noising Using a Fast Non-Local Averaging Algorithm

Permutation Correction in the Frequency Domain in Blind Separation of Speech Mixtures

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming

SOURCE separation techniques aim to extract independent

Robust Speaker Recognition using Microphone Arrays

Introduction to Blind Signal Processing: Problems and Applications

WARPED FILTER DESIGN FOR THE BODY MODELING AND SOUND SYNTHESIS OF STRING INSTRUMENTS

Speech Synthesis using Mel-Cepstral Coefficient Feature

Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks

A Wiener Filter Approach to Microphone Leakage Reduction in Close-Microphone Applications

Separation and Recognition of multiple sound source using Pulsed Neuron Model

29th TONMEISTERTAGUNG VDT INTERNATIONAL CONVENTION, November 2016

Blind Separation of Radio Signals Fading Channels

ROBUST echo cancellation requires a method for adjusting

Adaptive Speech Enhancement Using Partial Differential Equations and Back Propagation Neural Networks

FREQUENCY RESPONSE AND LATENCY OF MEMS MICROPHONES: THEORY AND PRACTICE

6-channel recording/reproduction system for 3-dimensional auralization of sound fields

Deblending random seismic sources via independent component analysis

Evaluation of a Multiple versus a Single Reference MIMO ANC Algorithm on Dornier 328 Test Data Set

Advanced delay-and-sum beamformer with deep neural network

Airo Interantional Research Journal September, 2013 Volume II, ISSN:

RECENTLY, there has been an increasing interest in noisy

Title. Author(s)Sugiyama, Akihiko; Kato, Masanori; Serizawa, Masahir. Issue Date Doc URL. Type. Note. File Information

The Steering for Distance Perception with Reflective Audio Spot

comes from recording each source separately in a real environment as described later Providing methodologies together with data sets makes it possible

Leak Energy Based Missing Feature Mask Generation for ICA and GSS and Its Evaluation with Simultaneous Speech Recognition

Correlated postfiltering and mutual information in pseudoanechoic model based blind source separation

Wavelet Speech Enhancement based on the Teager Energy Operator

Omnidirectional Sound Source Tracking Based on Sequential Updating Histogram

Transcription:

% > SEPARATION AND DEREVERBERATION PERFORMANCE OF FREQUENCY DOMAIN BLIND SOURCE SEPARATION Ryo Mukai Shoko Araki Shoji Makino NTT Communication Science Laboratories 2-4 Hikaridai, Seika-cho, Soraku-gun, yoto 619-2, Japan ryo,shoko,maki @cslab.kecl.ntt.co.jp ABSTRACT In this paper, we investigate the separation and dereverberation performance of frequency domain Blind Source Separation (BSS) based on Independent Component Analysis (ICA) by measuring impulse responses of a system. Since ICA is a statistical method, i.e., it only attempts to make outputs independent, it is not easy to predict what is going on in a BSS system physically. We therefore investigate the detailed components in the processed signals of a whole BSS system from a physical and acoustical viewpoint. In particular, we focus on the direct sound and reverberation in the target and jammer signals. As a result, we reveal that the direct sound of a jammer can be removed and the reverberation of the jammer can be reduced to some degree by BSS, while the reverberation of the target cannot be reduced. Moreover, we show that a long causes pre-echo noise, and this damages the quality of the separated signal. 1. INTRODUCTION Blind Source Separation (BSS) is a technique that separates and extracts target signals only from mixture signals observed without using information on the characteristics of the source signals and the acoustic system [1, 2]. Most BSS algorithms are considerably effective for instantaneous (non-convolutive) mixtures of signals, and some attempts have been made to apply BSS to signals mixed in convolutive environments [, 4]. However, it has also been pointed out that a sufficient performance cannot be obtained in environments with long reverberation where the filter lengths of the mixing and unmixing systems are on the order of thousands or higher [5, 6]. In this paper, we examine the performance of a separation system obtained by frequency domain BSS. We focus our attention on the power of (1) the direct sound of the target signal, (2) the reverberation of the target signal, () the direct sound of the jammer signal, and (4) the reverberation of the jammer signal, and evaluate each power separately. As a result, it is shown that frequency domain BSS based on ICA can remove the direct sound and reduce the reverberation of the jammer signal, while it hardly ever reduces the reverberation of the target signal. 2. FREQUENCY DOMAIN BSS OF CONVOLUTIVE MIXTURES When the source signals are ), the signals observed by microphone are, and the unmixed signals are, the model can be described by the following equations:!" $# /# '&)( 1&2(4 -,. (1) 5, 5 6 (2) where is the impulse response from source to microphone, 5 is the coefficient when the unmixing system is assumed as an FIR filter, and the operator, denotes convolution. In this paper, we consider a two-input, two-output convolutive BSS problem, i.e., = = 2 (Fig. 1). In addition, it is assumed that 8( 5 is separated to 9( 5, and :; is separated to 8:<. Because it is possible to convert a convolutive mixture in the time domain into an instantaneous mixture in the frequency domain, frequency domain BSS is effective for separating signals mixed in a reverberant environment. Using a = -point short-time Fourier transform for (1), we obtain > 5?@61AB $#BCD 5?E FG?H6AI J () We assume that the following separation has been completed in a frequency bin? :?H61AB $#?G?H6AI 6 (4) 2

> ] N source signals s1 h21 h11 observed signals x1 w21 w11 unmixed signals y1 (a) (b) (c) h11 h11 w11 s1 x1 s1 y1 h21 w12 h12 w11 w12 y1 h12 s2 h22 mixing system w12 x2 w22 unmixing system y2.8 x1s1 s2 h22 Fig. 1. Model of BSS system 5?@61AB # L M (?H6AI 6M : 5?@61AB ONQP is the 5?@61AB # LSR-(T?@61AB 16UR4:;?H61AB QNOP is the estimated source signal, and where observed signal in frequency bin?,?e represents the unmixing matrix. 5?E is determined so that R!(?H6AI and R<:; 5?@61AB become mutually independent. The above calculations are carried out for each frequency independently. For the calculation of unmixing matrix, we use an optimization algorithm based on the minimization of the ullback-leibler divergence [, 8]. The optimal is obtained by using the following iterative equation: -W (5) where Y a 'V)( # X L diag Y[ZH B\^] -_`Y ZH \^] denotes the averaging operator, is used to express the value of the -th step in the iterations, and X is the step size parameter. In addition, we define the nonlinear function ZH ä as ZH b# WIced Ref hg W Wiced Imf jg (6) where Re and Im are the real and imaginary parts of, respectively. In general, it is necessary to solve the permutation problem and scaling problem when ICA is used. In our experiment, the effect of the permutation problem was negligible and so we did not coordinate the permutation. The problem of scaling was solved by adjusting the power of the target signal in the output signal to db.. EVALUATION METHOD The performance of BSS is usually evaluated by the ratio of a target-originated signal to a jammer-originated signal. This measure is reasonable for evaluating the separation performance, but is unsuitable for evaluating the dereverberation performance because of its inability to distinguish the direct sound and reverberation. Since we want to know the detailed components in separated signals, i.e., the direct sound and reverberation of the target and jammer, we take the following procedure, -.8 (db) -1-2 - -4 PIR PTR PJ 125(ms) 125(ms) 125(ms) Fig. 2. Definitions of measurement factors. (1) estimate unmixing matrix?g for each frequency. (2) by using IFFT, transform frequency domain unmixing matrix?g to time domain unmixing filter 5. () while driving with the impulse as a source signal, measure four impulse responses, from ( to (, ( to :, : to (, and : to :. (4) investigate the four impulse responses in detail and compare them to the responses of a null beamformer (NBF)..1. Definitions of performance measurement factors We evaluate the performance of unmixing system in time domain. We consider a separated signal (, target signal (, and jammer signal :. When the target ( is an impulse k< and the jammer : #l, we call the observed signal ( as (nmt( [Fig. 2(a)], and ( as (nmo( [Fig. 2(b)]. Similarly, when ( #l and : #pk4 5, we call ( as (nmq:, and ( as (nm1: [Fig. 2(c)]. (rmo( is an impulse response from ( to ( by the mixing system C, and (nmt( is an impulse response from ( to ( by the whole system a5c. These are calculated by using and as follows. b(nmt(s# (nmq: # 9(nmo(u# (nm1: # W ([( () (t: (8) ([($, ([( (t:h, :"( (9) (t:vw : : (1) ([(, From the viewpoint of source separation, we can consider (nmo( as the direct and reverberant sound of target (, and (nm1: as the remaining sound of jammer :. (t:, 21

P P 5. m NBF.6 (a) ICA (b).12 m 2.15 m 1.56 m 4 cm microphones (height : 1.5 m) 1.15 m 4 1.15 m loudspeakers (height : 1.5 m) -.6.6 (c) (d) room height : 2. m Fig.. Layout of the room used in experiments. Reverberation time = ms. To simplify the evaluation, we normalize so that the power of the observed signals (rmo( and (nm1: is equal to db, and make the following definitions (Fig. 2). wix-y[z : the power of the reverberant sound in (nmt(, wix z : the power of the reverberant sound in{(nmo(, wix- : the power of (rm1:. We also define the reduction of the reverberation of target signal } and the reduction of jammer signal P } as follows } P # _~ x z _ x-y[z (11) } # _ x J (12) 4. EXPERIMENTS In order to examine what is separated by an unmixing system based on ICA, and what remains as noise, we investigated impulse responses of a system. In frequency domain BSS, it has been confirmed that the separation performance changes according to the length of the frame [6], so we chose the and the frame shift as parameters. 4.1. Conditions for the experiments The layout of the room we used to measure the impulse responses of the mixing system C is shown in Fig.. The reverberation time of the room was ms, which corresponds to impulse response of 24 taps at 8 khz sampling rate. We used a two-element array with inter-element spacing of cm. The speech signals arrived from two directions, i.e., _. 8lƒ and 8l. The contribution of the direct sound of ([( and :e( was 6.6 db, and that of (t: and :[: was 5. db. Two sentences spoken by two male speakers selected from the ASJ continuous speech corpus for research were used as the source signals. The lengths of these mixed -.6 ms 125ms ms 125ms Fig. 4. Target and jammer impulse responses of NBF and ICA speech signals were about eight seconds each. We used the entire eight seconds of the mixed data for learning according to (5). In these experiments, we changed the = from 2 to 496 and investigated the performance for each condition. The sampling rate was 8 khz, and analysis window was a Hamming window. The frame shift was =~ o and =~ o 8, which correspond to double and 2 times oversampling. The number of iterations for (5) was 1, except when ˆ#ˆ=~ o and =p# 124, 248, and 496, where the iteration was stopped at,, and 2, respectively, because a deterioration of the performance was observed. 4.2. Experimental results Figures 4(a) and (c) show examples of impulse responses (nmo( and (nmq: of the unmixing system obtained by a null beamformer (NBF) that forms a steep null directivity pattern towards a jammer under the assumption of the jammer s direction being known. Figures 4(b) and (d) are results obtained by ICA. For the target signal, we can see that the reverberation passes the system in both cases (NBF and ICA) in Figs. 4(a) and (b). Figure 4(c) shows that the direct sound of the jammer is removed, but the reverberation is not removed by NBF, as expected. On the other hand, Fig. 4(d) indicates that ICA not only removes the direct sound, but also reduces the reverberation of the jammer. Figure 5 shows the relationship between the frame length = and the reduction ratios } and P } defined by (11) and (12). } P ( and } ( are } P and } when the tar- 22

reduction ratio (db) reduction ratio (db) reduction ratio (db) 2 (a) ICA, S=T/2 15 1 5 2 15 1 5 15 1 5 (b) ICA, S=T/2 (c) NBF RJ1 RT1 RT2 2 64 128 256 512 124 248 496 T RJ2 RT1 RT2 RJ1 RJ2 RT1 RT2 RJ1 RJ2 Fig. 5. Relationship between and reduction ratio..6 -.6.6 -.6.6 -.6.6 -.6 ms observed signal T=512 ms T=248 ms T=496 25ms 25ms ms x1s2 ( x2s2) 25ms 25ms.6 -.6.6 -.6.6 -.6.6 -.6 ms ms 25ms ms 25ms 25ms pre-echo ms Fig. 6. Jammer impulse response of BSS system 5. DISCUSSION 25ms get signal is <(. } P : and } : are results when the target signal is :. Figures 5(a) and (b) show results by ICA when #Š=~ o and # =~ o 8, respectively. For the sake of comparison, the performance of NBF is shown in Fig. 5(c). Note that these results are measured by the power of impulse responses, and differ from the noise reduction rate (NRR) [6] measured by using a speech signal having a highly colored spectrum. Our results indicate seemingly better values than the NRR of the speech signal. For example, the reduction ratio } ( = 15.8 db and } : = 12.6 db (=Œ# 8l8 8Žƒ6U #I=~ o ) correspond to about 11 db and 8 db in the case of NRR, and } ( = 19.5 db and } : = 16.6 db (= # 8l8 8Žƒ6U #Œ=~ o 8 ) correspond to about 14 db and 9 db of NRR. First, we discuss the jammer reduction ratio }. When = 8Ž, the reduction performance of BSS is as poor as that of NBF, and when 8 8 =i 8l8 8Ž, the reduction ratio increases. In the case of =p# 8l8 Ž, ˆ# =~ o 8, } ( =19.5 db, } : =16.6 db. This is greater than the contribution of the direct sound, i.e., 6.6 db and 5. db. This means that the unmixing system by ICA can reduce not only the direct sound of the jammer but also the reverberant sound of the jammer. In addition, comparing the results of # =~ o and D#=^ T 8 [Figs. 5(a) and (b)], we can see that oversampling improves the jammer reduction ratio. However, as we describe later, the reverberation is not eliminated completely. On the other hand, the reduction ratio of the reverberation of target } is low, and does not vary through the P 2

.6 -.6.6 -.6.6 ms observed signal T=512 ms T=248 25ms 25ms x1s1 ( x2s1).6 -.6.6 -.6.6 ms 25ms ms 25ms = 512, 248, and 496. The best performance is obtained when = = 248. In the case of = = 512, the length of the unmixing system is much shorter than the length of the reverberation; accordingly, the reverberation longer than the frame cannot be reduced at all. On the other hand, when = = 496, which is longer than the reverberant time, the unmixing system can wholly cover the reverberation, but because each tap of the filter has errors that derive from the statistical method of ICA. When the filter length becomes longer, the number of coefficients to be estimated increases while the number of data for learning in each frequency bin decreases. As a result, the amount of estimation errors escalates. Moreover, the pre-echo noise grows, and this causes the performance to fall. The target signal s impulse response -(rmo( is shown in Fig.. As we have described previously, the reverberation is not removed. Furthermore, the target signal still suffers from a pre-echo noise, and this damages the quality of the separated target signal. -.6.6 -.6 ms T=496 ms 25ms 25ms -.6.6 -.6 ms pre-echo ms 25ms Fig.. Target impulse response of BSS system 25ms entire =. This means that dereverberation was not achieved for the target signal. From these results, it can be concluded that is not the approximation of the inverse system of C, but a filter that can eliminate the jammer signal. It has been pointed out that early reflections of the jammer signal are removed by BSS [9]. We obtained a slightly stronger result that not only the early reflections but also the reverberation of the jammer signal is reduced to some degree. The reason for this is that frequency domain BSS is equivalent to two sets of frequency domain adaptive microphone arrays, i.e., Adaptive Beamformers (ABF), which adapt to minimize the jammer signal including reverberation in the mean square error sense [1]. Finally, we show the reason why the reduction ratio of jammer signal } declines when = is too long. Figure 6 shows the jammer signal s impulse response (nm1:, when = 6. CONCLUSION We investigated the performance of an unmixing system obtained by frequency domain BSS based on ICA using the impulse responses of target and jammer signals. As a result, we revealed that ICA not only removes the direct sound of the jammer signal, but also reduces the reverberation, while the reverberation of the target is not reduced. The jammer reduction performance increases as the becomes longer. However, an overly long decreases the performance due to accumulating errors. The performance of the target dereverberation does not depend on the and is as poor as that of NBF. ACNOWLEDGEMENTS We would like to thank Dr. Hiroshi Saruwatari for his valuable discussions. We also thank Dr. Shigeru atagiri for his continuous encouragement. REFERENCES [1] A. J. Bell and T. J. Sejnowski, An informationmaximization approach to blind separation and blind deconvolution, Neural Computation, vol., no. 6, pp. 1129 1159, 1995. [2] S. Haykin, Ed., Unsupervised adaptive filtering, John Wiley & Sons, 2. 24

[] T. W. Lee, A. J. Bell, and R. Orglmeister, Blind source separation of real world signals, Neural Networks, vol. 4, pp. 2129 214, 199. [4] J. Xi and J. P. Reilly, Blind separation and restoration of signals mixed in convolutive environment, in Proc. ICASSP 9, 199, pp. 12 1. [5] M. Z. Ikram and D. R. Morgan, Exploring permutation inconsistency in blind separation of speech signals in a reverberant environment, in Proc. ICASSP, 2, pp. 141 144. [6] S. Araki, S. Makino, T. Nishikawa, and H. Saruwatari, Fundamental limitation of frequency domain blind source separation for convolutive mixture of speech, in Proc. ICASSP21, 21, MULT-P2.1. [] S. Ikeda and N. Murata, An informationmaximization approach to blind separation and blind deconvolution, in Proc. ICA99, 1999, pp. 65. [8] S. urita, H. Saruwatari, S. ajita,. Takeda, and F. Itakura, Evaluation of blind signal separation method using directivity pattern under reverberant conditions, in Proc. ICASSP, 2, pp. 14 14. [9] F. Asano, S. Ikeda, M. Ogawa, H. Asoh, and N. itawaki, Blind source separation in reflective sound fields, in Proc. Int. Workshop on Hans-Free Speech Communication 21, 21, pp. 51 54. [1] S. Araki, S. Makino, R. Mukai, and H. Saruwatari, Equivalence between frequency domain blind source separation and frequency domain adaptive null beamformers, in Proc. Eurospeech21, 21. 25