2 ieee transactions on audio, speech, and...

11
2 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING Robust Source Localization and Enhancement With a Probabilistic Steered Response Power Model Johannes Traa* Student Member, IEEE, David Wingate, Noah D. Stein, Paris Smaragdis Fellow, IEEE Abstract—Source localization and enhancement are often treated separately in the array processing literature. One can ap- ply Steered Response Power (SRP) localization to determine the sources’ Directions-Of-Arrival (DOA) followed by beamforming and Wiener post-filtering to isolate the sources from each other and ambient interference. We show that when there is significant overlap between directional sources of interest in the Time- Frequency (TF) plane, traditional SRP localization breaks down. This may occur, for example, when the array is located near a reflector, significant early reflections are present, or the sources are harmonized. We propose a joint solution to the localization and enhancement problems via a probabilistic interpretation of the SRP function. We formulate optimization procedures for (1) a mixture of single-source SRP distributions (MoSRP) and (2) a multi-source SRP distribution (MultSRP). Unlike in traditional localization, the latter approach explicitly models source overlap in the TF plane. Results shows that the MultSRP model is capable of localizing sources with significant overlap in the TF domain and that either of the proposed methods out-performs standard SRP localization for multiple speakers. Index Terms—steered response power, source localization, beamforming, blind source separation I. I NTRODUCTION The array processing literature contains methods [1], [2] for enhancing directional signals in the presence of interferers and noise. Beamforming [3] aims to optimize a spatial filter that isolates a target signal’s energy. The simplest example is the Delay-and-Sum (DS) beamformer, which is optimal for a single source in diffuse, additive, white Gaussian noise. This is a data-independent method since the true noise character- istics don’t enter into the design of the filter. For general Gaussian noise, the optimal solution is the Minimum-Variance Distortionless Response (MVDR) beamformer. Finally, when several targets and/or interferers are present simultaneously, the Linearly-Constrained Minimum-Variance (LCMV) beam- former [4] provides a means of implementing both distortion- less and null constraints. The former protect the target signal while the latter block undesired signals. This assumes that each source’s Direction-Of-Arrival (DOA) is known. So, in practice, beamforming is preceded by a localization step. Beamforming is a linear processor and so has limited source separation capabilities. Combining a beamformer with Manuscript received ... ..., 2015. J. Traa is a PhD student in the ECE department at the University of Illinois at Urbana-Champaign (UIUC) ([email protected]). D. Wingate and N. Stein are researchers at Lyric Labs, Analog Devices ([email protected], [email protected]). P. Smaragdis holds a joint faculty position in ECE and CS at UIUC and works with Adobe Systems, Inc. ([email protected]) Wiener post-filtering [5] has been shown to provide adap- tive noise reduction for non-stationary signals. However, the source DOAs are still assumed to be known. Some blind beamforming methods ignore the array structure and source DOAs altogether. The Joint Approximate Diagonalization of Eigen-matrices (JADE) algorithm was applied to recover the mixing matrix that describes the channel between sources and sensors [6]. There are many methods for localizing (and tracking) di- rectional sources [7], [8], [9], [10], [11]. Steered Response Power (SRP) localization [12] involves computing the output power of a beamformer steered towards each DOA of interest and locating one or more peaks in the resulting SRP function. A similar approach computes a Generalized Cross-Correlation (GCC) [13] function over time delays. Spectral weighting such as the Phase Transform (PHAT) were explored in [13] for enhancing the SPR and GCC functions. The authors in [12], [14] used the GCC-PHAT approach to efficiently localize speech sources on a hemisphere. A third method centers around the eigenanalysis of the channel correlation matrix. The Multiple Signal Classification (MUSIC) algorithm [15] iden- tifies signal and noise subspaces to form a “pseudo-spectrum” that contains peaks at the source DOAs. This requires a scan over DOA space. The root-MUSIC algorithm [16] avoids this by reducing the localization problem to one of root-finding. A related method, ESPRIT [17], involves a similar analysis, but takes advantage of arrays with a special structure. Several authors have extended these eigen-analysis methods to handle wideband sources [18]. Several authors have described the relationship between the GCC and SRP functions and a probabilistic model of the observed signals [12], [19], [20]. Similarly, SRP local- ization has been described as a Maximum-Likelihood (ML) problem [21], [22]. In [23], the authors model the observed frequency-domain data vectors as zero-mean Gaussian random variables and use an EM algorithm to learn the covariance parameters of the sources and apply multichannel Wiener filtering to perform source separation. The authors in [24] formulate a DOA-dependent covariance matrix to localize a single source in noisy conditions. Time-frequency (TF) masking [25] is known to outperform linear filtering (e.g. beamforming) methods for enhancing wideband sources. This approach was originally motivated by the disjointness of speech over the TF plane [26]. The trade-off is the introduction of musical noise artifacts in the enhanced/separated signals. Despite this, much work has been done on source localization and separation using TF fea- tures [27], [28]. In [29], the authors localize the sources with

Upload: doanbao

Post on 14-Jun-2018

218 views

Category:

Documents


1 download

TRANSCRIPT

2 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING

Robust Source Localization and Enhancement Witha Probabilistic Steered Response Power ModelJohannes Traa* Student Member, IEEE, David Wingate, Noah D. Stein, Paris Smaragdis Fellow, IEEE

Abstract—Source localization and enhancement are oftentreated separately in the array processing literature. One can ap-ply Steered Response Power (SRP) localization to determine thesources’ Directions-Of-Arrival (DOA) followed by beamformingand Wiener post-filtering to isolate the sources from each otherand ambient interference. We show that when there is significantoverlap between directional sources of interest in the Time-Frequency (TF) plane, traditional SRP localization breaks down.This may occur, for example, when the array is located near areflector, significant early reflections are present, or the sourcesare harmonized. We propose a joint solution to the localizationand enhancement problems via a probabilistic interpretation ofthe SRP function. We formulate optimization procedures for (1)a mixture of single-source SRP distributions (MoSRP) and (2) amulti-source SRP distribution (MultSRP). Unlike in traditionallocalization, the latter approach explicitly models source overlapin the TF plane. Results shows that the MultSRP model is capableof localizing sources with significant overlap in the TF domainand that either of the proposed methods out-performs standardSRP localization for multiple speakers.

Index Terms—steered response power, source localization,beamforming, blind source separation

I. INTRODUCTION

The array processing literature contains methods [1], [2]for enhancing directional signals in the presence of interferersand noise. Beamforming [3] aims to optimize a spatial filterthat isolates a target signal’s energy. The simplest example isthe Delay-and-Sum (DS) beamformer, which is optimal for asingle source in diffuse, additive, white Gaussian noise. Thisis a data-independent method since the true noise character-istics don’t enter into the design of the filter. For generalGaussian noise, the optimal solution is the Minimum-VarianceDistortionless Response (MVDR) beamformer. Finally, whenseveral targets and/or interferers are present simultaneously,the Linearly-Constrained Minimum-Variance (LCMV) beam-former [4] provides a means of implementing both distortion-less and null constraints. The former protect the target signalwhile the latter block undesired signals. This assumes thateach source’s Direction-Of-Arrival (DOA) is known. So, inpractice, beamforming is preceded by a localization step.

Beamforming is a linear processor and so has limitedsource separation capabilities. Combining a beamformer with

Manuscript received ... ..., 2015.J. Traa is a PhD student in the ECE department at the University of Illinois

at Urbana-Champaign (UIUC) ([email protected]).D. Wingate and N. Stein are researchers at Lyric Labs, Analog Devices

([email protected], [email protected]).P. Smaragdis holds a joint faculty position in ECE and CS at UIUC and

works with Adobe Systems, Inc. ([email protected])

Wiener post-filtering [5] has been shown to provide adap-tive noise reduction for non-stationary signals. However, thesource DOAs are still assumed to be known. Some blindbeamforming methods ignore the array structure and sourceDOAs altogether. The Joint Approximate Diagonalization ofEigen-matrices (JADE) algorithm was applied to recover themixing matrix that describes the channel between sources andsensors [6].

There are many methods for localizing (and tracking) di-rectional sources [7], [8], [9], [10], [11]. Steered ResponsePower (SRP) localization [12] involves computing the outputpower of a beamformer steered towards each DOA of interestand locating one or more peaks in the resulting SRP function.A similar approach computes a Generalized Cross-Correlation(GCC) [13] function over time delays. Spectral weighting suchas the Phase Transform (PHAT) were explored in [13] forenhancing the SPR and GCC functions. The authors in [12],[14] used the GCC-PHAT approach to efficiently localizespeech sources on a hemisphere. A third method centersaround the eigenanalysis of the channel correlation matrix. TheMultiple Signal Classification (MUSIC) algorithm [15] iden-tifies signal and noise subspaces to form a “pseudo-spectrum”that contains peaks at the source DOAs. This requires a scanover DOA space. The root-MUSIC algorithm [16] avoids thisby reducing the localization problem to one of root-finding.A related method, ESPRIT [17], involves a similar analysis,but takes advantage of arrays with a special structure. Severalauthors have extended these eigen-analysis methods to handlewideband sources [18].

Several authors have described the relationship betweenthe GCC and SRP functions and a probabilistic model ofthe observed signals [12], [19], [20]. Similarly, SRP local-ization has been described as a Maximum-Likelihood (ML)problem [21], [22]. In [23], the authors model the observedfrequency-domain data vectors as zero-mean Gaussian randomvariables and use an EM algorithm to learn the covarianceparameters of the sources and apply multichannel Wienerfiltering to perform source separation. The authors in [24]formulate a DOA-dependent covariance matrix to localize asingle source in noisy conditions.

Time-frequency (TF) masking [25] is known to outperformlinear filtering (e.g. beamforming) methods for enhancingwideband sources. This approach was originally motivatedby the disjointness of speech over the TF plane [26]. Thetrade-off is the introduction of musical noise artifacts in theenhanced/separated signals. Despite this, much work has beendone on source localization and separation using TF fea-tures [27], [28]. In [29], the authors localize the sources with

3

X

(w)

TF#Weigh)ng#(Beamformer#+#Wiener#

Mask,#Posteriors)#

DOA#Update#

Correla)on#Matrices#

R

Fig. 1: Proposed iterative approach. A probabilistic SRPmodel is combined with time-frequency masking to performblind source localization and separation in the presence ofnon-directional interference. Observed data X collected withmicrophones is used to optimize source DOAs ⇥ usingchannel correlation information R, Wiener mask weights �,and possibly source-specific TF weights w.

visual cues and separate them by applying data-independentbeamforming followed by TF masking. And in [30], a TFmasking approach is proposed for tracking acoustic sourceswith a Rao-Blackwellized particle filter.

In this paper, we interpret the SRP function as a likelihoodand propose two methods for maximizing it as a function of thesource DOAs. One uses a mixture of single-source SRPs andthe other uses an SRP that explicitly models the presence ofmultiple sources. We show that the latter is robust to situationswhere the sources have a significant amount of overlap in theTF plane. This occurs when, for example, early reflections arepresent [31] (e.g. the array is located near a reflective surface)or the sources are harmonized as when instrumentalists playin unison. We apply TF masking [25], [27] to isolate TFbins that correspond to the directional signals of interest. Inthis way, the localization, separation, and Wiener post-filteringsteps are merged into one. The flow diagram for the proposedapproach is shown in Fig. 1. This work is motivated in partby unified approaches that have been shown to out-performsequential approaches for automatic speech recognition [32].Indeed, we show that the proposed method outperforms astandard sequential localization approach.

The contributions of this paper are:• A discussion of the probabilistic interpretation of the

Steered Response Power (SRP) function.• An SRP model that explicitly considers the presence of

multiple simultaneous sources.• A joint, iterative method for source localization and

enhancement/separation applied to a mixture of SRPsmodel and the proposed multi-source model.

• Experiments demonstrating the benefits of the proposedapproach for source localization and separation.

II. DATA MODEL

Consider the convolutive time-domain model of a singlesource recorded under noisy conditions at M channels:

xm [t] = am [t] ⇤ s [t] + nm [t] , (1)

where xm [t] is the recorded sample at the m

th channel,am [t] is the Room Impulse Response (RIR) between thesource and m

th channel, nm [t] is a Gaussian noise process,and ⇤ denotes convolution. We apply the Short-Time FourierTransform (STFT) to the mixed, noisy signal xm to de-couplethe signal components across frequency and time. Thus, attime frame t 2 [1, T ] and frequency f 2 [1, F ], we have:

xft = af sft + nft , (2)

where xft 2 CM is an observed data vector, af 2 CM isthe mixing vector, sft 2 C is the source coefficient, andnft 2 CM contains the noise coefficients. For simplicity, weassume that nft ⇠ N

0,�

2fI

and that the source and noisecoefficients are statistically independent.

For a point source in an anechoic environment, we writeaf explicitly in terms of the source DOA as the unit steeringvector:

af (�) =

1pM

exp

j

2⇡lf

c

m

>�

, (3)

where � 2 R3 is the source’s unit DOA vector, m 2 R3⇥M

is the matrix of M sensor positions, lf is the center frequencyof the f

th band, and c is the speed of sound.When K > 1 sources are present, we can write:

xft = Af (�) sft + nft , (4)

where � = {�1:K} denotes the set of source DOAs, sft 2 CK

is a vector of source coefficients, and we have defined thesteering matrix:

Af (�) =

af (�1) · · · af (�K)

2 CM⇥K. (5)

III. CLASSICAL ARRAY PROCESSING

In this section, we describe how the propagation modelsin (2) and (4) are used for narrowband beamforming and lo-calization. We can easily extend this to the broadband case byassuming that all frequency bands are mutually independent.

A. BeamformingLinear spatial filters are often used for enhancing directional

signals. They can be described by a weight vector wf 2 CM

that is used to estimate a source coefficient sft via bsft =

w

Hf xft. The optimal wf minimizes the expected output power

of the filter:

P = E

|bsft|2⇤

= w

Hf Rf wf , (6)

without distorting the desired signal at DOA �:

b

wf = argmin

wf

w

Hf Rf wf , (7)

s.t. a

Hf (�)wf = 1 . (8)

where we have defined the channel correlation matrix:

Rf = E

xft xHft

. (9)

The solution is the well-known MVDR beamformer:

4 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING

b

w

MVDRf =

R

�1f af (�)

a

Hf (�)R

�1f af (�)

. (10)

If we assume (in the derivation) that Rf = I, this reducesto the data-independent Delay-and-Sum (DS) beamformer:

b

w

DSf = af (�) , (11)

with corresponding empirical output power:

PDS = a

Hf (�)

b

Rf af (�) , (12)

where:

b

Rf =

1

T

TX

t=1

xft xHft , (13)

is an empirical approximation of (9). The DS beamformercan be pre-computed, while the MVDR beamformer providesbetter noise suppression when Rf 6= I. They are generalizedto the multi-source case via the multiply-constrained optimiza-tion problem:

b

wf = argmin

wf

w

Hf Rf wf , (14)

s.t. A

Hf (�)wf = u , (15)

where u 2 CK contains the desired gains for all DOAs. Thesolution is sometimes refered to as the Linearly-ConstrainedMinimum-Variance (LCMV) beamformer [4]:

b

w

LCMVf = R

�1f Af (�)

h

A

Hf (�)R

�1f Af (�)

i�1u . (16)

B. Steered Response Power (SRP) LocalizationBeamformers can be used if the source DOA(s) � are

known. SRP localization [12] aims to identify � by searchingfor peaks in the output power of a single-source beamformer athypothesized DOAs ✓. We can see this for the DS beamformerand one source by writing (12) as:

PDS (✓) =

1

T

TX

t=1

|aHf (✓)xft|2 (17)

=

1

T

TX

t=1

|aHf (✓)af (�) |2 |sft|2 + C(f, T ) (18)

1

T

TX

t=1

|sft|2 + C(f, T ) , (19)

where C(f, T )

T!1����! �

2f is due to the additive noise term and

equality holds when ✓ = �. The Phase Transform (PHAT) [13]involves setting all the magnitudes of the components of xft

to 1 and helps to emphasize the peaks in PDS (✓).When multiple sources are present, one might search for

multiple peaks in the SRP [12]. There are three issues withthis approach. First, this scan over DOA space may beinfeasible when computation power is limited or we desirea high spatial resolution. Efficient strategies for searchingfor peaks in the SRP function have been proposed to work

around this issue [33], [34]. Second, we expect the peaksto be poorly localized at low frequencies and for closely-spaced microphones because the steering vectors are hard todistinguish under those conditions. And third, if the sourcecoefficients are simultaneously large in magnitude, the SRPfunction is distorted by cross-terms.

A more effective approach scans over all DOA sets �

using an LCMV beamformer and locates the peak outputpower. However, this leads to a combinatorial problem. If wediscretize the DOA search space into D look directions, wemust scan over D

K�’s. We remedy this by modeling the

multi-source SRP function as a continuous likelihood functionparametrized by � and maximizing it with gradient ascent.

IV. PROBABILISTIC SRP MODEL

We present a probabilistic model [12] for the observed datavectors xft and relate this to the SRP function described inSection III-B. A more detailed account of this probabilisticformulation can be found in Appendix A.

A. SRP Likelihood

The propagation model in (4) corresponds to a Gaussianlikelihood for the observed data vectors:

log Lft (⇥) = log N�

xft ; µft , �2f I

. (20)

The mean µft encodes the expected value of xft:

µft = E [xft] = Af (⇥) E [sft|xft] , (21)

for a hypothesized DOA set ⇥. We approximate the expecta-tion with a least-squares estimate:

b

sft =

A

Hf (⇥)Af (⇥)

⇤�1A

Hf (⇥) xft , (22)

which we recognize as the output of a data-independentLCMV beamformer (i.e. Rf / I). Thus, we have:

bµft = Af (⇥)

b

sft = Bf (⇥) xft , (23)

where:

Bf (⇥) = Af (⇥)

A

Hf (⇥)Af (⇥)

⇤�1A

Hf (⇥) , (24)

is a projection matrix. Now we can write (20) as a zero-meanGaussian likelihood:

log Lft (⇥) / � 1

2�

2f

x

Hft Pf (⇥)xft , (25)

in terms of a precision matrix:

Pf (⇥) = I�Bf (⇥) . (26)

Aggregating over all t and expanding, we have:

log Lf (⇥) / � 1

2�

2f

TX

t=1

kxftk22 � x

Hft Bf (⇥) xft , (27)

which, in the one-source case, simplifies to:

5

log Lf (✓) / � 1

2�

2f

TX

t=1

kxftk22 � |aHf (✓) xft|2 . (28)

This is equivalent to the SRP function defined in (17) asfar as identifying the true DOA is concerned. As in [12],we can view beamforming-based localization as a maximum-likelihood problem.

B. Effect of Cross-Talk

To show the effect of cross-talk on the SRP functions, wegenerated a mixed data vector from two directional sourcesof comparable magnitude. We then computed the single-and multi-source likelihoods as well as a product of single-source likelihoods on the unit semicircle as shown in Fig. 2.Interference between the data vectors results in a spurious peakin the single-source likelihoods about halfway between thetrue DOAs. In contrast, the DOAs are correctly identified inthe two-source likelihood up to a labeling permutation.

We can also investigate the effect of cross-talk mathemati-cally. Consider the case of K = 2 sources. We can write themain term of the multi-source likelihood as:

pMultSRP (xft) = x

HftBxft

=

kAHxftk22 � 2 Re

n

a

H2 a1

� �

a

H1 xft

x

Hfta2

⌘o

1� |aH2 a1|2

. (29)

If we were to model the data with a sum of one-sourcemodels, we would have:

pMoSRP (xft) =

2X

k=1

x

Hftaka

Hk xft = kAH

xftk22 . (30)

We can see that additional terms are present in the multi-source model that can change the location of the peak in theSRP function. This is observed to occur when the sources co-activate in the time-frequency plane. As an example, considerthe case of two sources with equal loudness and oppositeangles on the unit circle, i.e. xft = A1 = a (✓)+a (�✓). Themulti-source model would exhibit a peak at the true DOA pair.Meanwhile, the single-source model would exhibit a peak ata DOA ✓0 6= ±✓ such that |aH

✓0�xft|2 is maximized. A

similar case is shown in Fig. 2.

V. MAXIMUM-LIKELIHOOD LOCALIZATION

We seek a maximum-likelihood estimate of the sourceDOAs. We will do this by gradient ascent on the SRPlikelihood (27):

i ⇥

i�1+ ⌘i ⌦

0

@

FX

f=1

@ logLf (⇥)

@ ⇥

i�1

1

A

, (31)

where ⌦ (�) performs column-wise normalization and ⌘i =

⌘0 (Imax � i) /Imax is a decaying step size.

0 0.5 1 1.5 2 2.5 3

−0.06

−0.05

−0.04

−0.03

DOA (radians)

L

ike

lih

oo

d

Single−Source Likelihood

0 1 2 30

0.5

1

1.5

2

2.5

3

Source 1 DOA (radians)

S

ou

rce

2 D

OA

(ra

dia

ns

)

Sum of Single−Source Likelihoods

−0.13

−0.12

−0.11

−0.1

−0.09

−0.08

−0.07

−0.06

0 1 2 30

0.5

1

1.5

2

2.5

3

Source 1 DOA (radians)

S

ou

rce

2 D

OA

(ra

dia

ns

)

Multi−Source Likelihood

−0.03

−0.025

−0.02

−0.015

−0.01

−0.005

Fig. 2: SRP likelihoods for a toy data mixture of two sourceswith DOAs at 0.56 and 2.26 radians on the unit circle.Squares and dashed lines indicate their maxima. (Top) Single-source likelihood. (Middle) Product of single-source likeli-hoods. (Bottom) Multi-source likelihood.

A. One-Source Model

In the one-source case, the gradient can be written as:

@ logLf (✓)

@ ✓=

1

2f

2⇡lf

c

m Im⇥

a

⇤f (✓)� (Rfaf (✓))

, (32)

where � denotes element-wise multiplication and ⇤ denotescomplex conjugation. When multiple sources are present, wemodel the presence of the sources at each time t with hiddenvariables zft that capture which source is active. We iteratebetween estimating the zft’s and the DOAs in a GeneralizedEM1 (GEM) algorithm [36] corresponding to the completedata log likelihood:

1GEM is a variant of EM [35] that only requires the M step to increasethe lower bound rather than explicitly maximize it.

6 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING

Algorithm 1 Robust MoSRP LocalizationPreprocessing

Compute STFT matrices X

(k) 2 CF⇥T

Apply PHAT weighting: ˜

X � X ./ |X|Initialize DOAs: ✓(k) � ✓(k)

0

Optimizationfor i = 0 : Imax � 1 do

Compute steering matrices: Af =

1pMe

j2⇡lf

c m

>⇥

Compute projectors: Bf = Af

h

A

Hf Af

i�1A

Hf

if Wiener mask thenEstimate TF weights: b�ft =

kBfxftk22

kxftk22

elseNo weighting: b�ft = 1

end ifCompute posteriors: log w

(k)ft /

|a(k)f

Hx̃ft|2

2�2f

Estimate correlation matrices:b

R

(k)f =

TP

t=1b�ft w

(k)ft ˜

xft˜xHft

Compute gradients: g

(k)=

FP

f=1lf · · ·

m Imh

a

(k)f

⇤�⇣

b

R

(k)f a

(k)f

⌘i

Normalize gradients: g

(k) � g

(k)/ kg(k)k2

Compute step size: ⌘i = ⌘0 (Imax � i) /Imax

Update DOA vectors: ✓(k) � ✓(k)+ ⌘i g

(k)

Project DOAs to hemisphere: ✓(k) � ✓(k)/k✓(k)k2

end for

log p (X,Z |⇥) /FX

f=1

TX

t=1

KX

k=1

1

2�

2f

[zft = k] |aHf (✓k)xft|2 .

(33)

where [�] is the indicator function. The EM lower bound canbe written as:

Q /FX

f=1

KX

k=1

1

2�

2f

a

Hf (✓k)

b

R

(k)f af (✓k) , (34)

where:

b

R

(k)f =

TP

t=1w

(k)ft xft x

Hft

TP

t=1w

(k)ft

, (35)

are source-specific correlation matrices defined in terms of theposterior probabilities of the zft’s:

logw

(k)ft = log p (zft = k |xft) /

a

Hf (✓k) xft

2

2�

2f

. (36)

In the E step, we compute soft TF weights and correlationmatrices with (35)-(36), and in the M step, we optimize each

Algorithm 2 Robust MultSRP LocalizationPreprocessing

Compute STFT matrices X

(k) 2 CF⇥T

Apply PHAT weighting: ˜

X � X ./ |X|

Estimate correlation matrices: bRf =

1N

TP

t=1˜

xft˜xHft

Initialize DOAs: ⇥ � ⇥0

Optimizationfor i = 0 : Imax � 1 do

Compute steering matrices: Af =

1pMe

j2⇡lf

c m

>⇥

Compute projectors: Bf = Af

h

A

Hf Af

i�1A

Hf

if Wiener mask thenEstimate TF weights: b�ft =

kBfxftk22

kxftk22

Estimate correlation matrices:b

Rf =

TP

t=1b�ft˜xft˜x

Hft

end ifCompute gradient: G =

FP

f=1lf · · ·

m Im

A

⇤f �

(I�Bf )

b

RfAf

h

A

Hf Af

i�1◆�

Normalize source gradients: g

(k) � g

(k)/ kg(k)k2

Compute step size: ⌘i = ⌘0 (Imax � i) /Imax

Update DOA matrix: ⇥ � ⇥ + ⌘i G

Project DOAs to hemisphere: ✓(k) � ✓(k)/k✓(k)k2

end for

source’s DOA with (31)-(32). Thus, the EM procedure alter-nates between estimating localization (DOA) and separation(TF mask) parameters to fit a Mixture of SRP functions(MoSRP). For simplicity, we assume that all the mixingweights are equal and use a constant variance of �

2f = 10

�2

when calculating posterior probabilities.

B. Multiple-Source ModelThe gradient for multiple sources is (see Appendix B):

@ logLf (⇥)

@ ⇥

=

1

2f

2⇡lf

c

· · ·

m Imh

A

⇤f �

PfRfAf

A

Hf Af

⇤�1⌘i

, (37)

Gradient ascent with this model avoids the complexity of theEM algorithm while explicitly accounting for cross-talk. Anefficient implementation for K = 2 sources is derived byexpanding (37) in terms of the source-specific steering vectors(see Appendix C).

C. Robustness to Non-Directional InterferenceWhen non-directional interference eft is present, we have:

xft = Af (�) sft + nft + eft = bft + cft , (38)

where bft = Af (�) sft and cft = nft + eft. The Wienermask [25] gives the MMSE-optimal weighting to recover bft:

7

optft =

kbftk22kbftk22 + kcftk22

. (39)

Thus, a robust estimate of the correlation matrices is:

b

Rf =

TP

t=1�

optft xft x

Hft

TP

t=1�

optft

. (40)

In practice, we approximate the weights using:

b

bft = Bf (⇥)xft , (41)

b

cft = xft � bbft , (42)

which gives:

b�ft =

kbbftk22kbbftk22 + kbcftk22

=

kbbftk22kxftk22

. (43)

Interleaving this step with DOA optimization improveslocalization accuracy in the presence of ambient noise. Fora mixture of one-source models, we estimate the correlationmatrices as in (35) by multiplying the posteriors with theWiener filter weights. Since the gradients are accumulatedover all frequency bands, we weight the contributions fromeach band by the sum of all the corresponding TF weights.Due to the form of the gradient, this is equivalent to omittingthe denominator in (40).

The overall iterative scheme is shown in Fig. 1. Pseudocodefor the Mixture of SRPs (MoSRP) and Multi-source SRP(MultSRP) localizers are given in Algorithms 1 and 2.

VI. SOURCE SEPARATION

Once the gradient ascent procedure has converged, anynumber of methods can be used to enhance/separate thedirectional signals, if necessary. For example, we can isolateeach source by estimating time-frequency (TF) masks andapplying them to X. The mask weights are calculated as:

logw

(k)ft /

1

2⌫

2| bs (k)

ft |2 , (44)

using estimates of the source coefficients provided by K data-independent LCMV beamformers, each designed to isolatea single source while blocking out the others. This can beimplemented as:

b

sft =

A

Hf Af

⇤�1A

Hf xft . (45)

The variance ⌫ 2 (0,1) controls the hardness of the masksuch that as ⌫ ! 0, the mask becomes binary, assigning eachTF bin entirely to a single source. In our experiments, we used⌫ = 0.1.

VII. EXPERIMENTS

A. Speaker Localization in Noisy ConditionsWe ran speaker localization experiments in a room simulator

of size 5⇥5⇥5 meters with an 8-channel, 10⇥10 centimetersquare array placed in the center of the room. The microphones

Fig. 3: Localization errors for white Gaussian noise and anideal initialization. All gradient methods perform well withthe multi-source models performing slightly better.

were equally spaced around the perimeter of the square(including the corners). Between 2 and 4 sources were placedon the unit hemisphere above the array at random, ensuringsufficient separation: 8 j 6= k cos

�1⇣

�>j �k

� 2⇡K+2 . For

each speaker, we concatenated random sentences from theTSP database [37] to form a 10-second-long source signal.Reverberation with a T60 time of 260 milliseconds was simu-lated with the image method [38]. Varying amounts of eitherof two types of ambient noise were added instantaneouslyto the reverberant mixtures: (1) white Gaussian noise and(2) crowd noises from the BBC Sounds Effects Library. Ineach simulation with crowd noise, one single-channel noisesignal was replicated across the channels and each replicatewas randomly circularly shifted in time to ensure spatialincoherence. All STFTs were computed with window and hopsizes of 1024 and 256 using a Hann window.

We compared three methods: mixture of one-source SRPs(MoSRP), multi-source SRP (MultSRP), and traditional SRPlocalization (“Sequential”). The first two were tested with andwithout the Wiener filtering stage described in Section V-C. Inthe figures, algorithms using Wiener filtering are labeled with“(W)”. In addition, we ran the MultSRP method initializedwith (i.e. with a warm start from) either the MoSRP methodor the sequential method (MultSRP (ws1) and MultSRP (ws2),respectively). The sequential method was implemented with agrid of 709 DOAs uniformly spread over the hemisphere.

We found that the initialization method is important forthe SRP models. To evaluate this, we ran experiments bothwith and without an ideal initialization. The ideal initializationinvolves placing the initial DOA vector estimates at the truesource DOAs. Our practical initialization method is based onfitting wrapped lines to Inter-channel Phase Difference (IPD)features with the Random Sample Consensus (RANSAC)algorithm [39]. A similar procedure was used in [23].

Localization errors averaged over 100 trials are shown inFigs. 3-6. The error is averaged over the sources and capturesangular deviation from the true DOAs:

e = min

P

1

K

KX

k=1

cos

�1⇣

�>kb✓P(k)

(46)

where P is a permutation mapping. Minimizing over permu-

8 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING

Fig. 4: Localization errors for crowd noise and an idealinitialization. All gradient methods perform well with themulti-source models performing slightly better.

Fig. 5: Localization errors for white Gaussian noise and apractical initialization. The MultSRP model performs bestwhen initialized with the MoSRP result.

tations deals with the source labeling ambiguity.We see that with an ideal initialization and either noise type,

the MultSRP algorithms consistently out-perform the others.This is probably due its robustness to noise once the correctDOAs are identified. We can understand this by observingthat the projection matrices Bf behave like a series of LCMVbeamformers steered towards the sources (see Algorithm 2). Incontrast, the MoSRP method effectively steers delay-and-sumbeamformers towards the sources, making it more sensitive tonoise and cross-talk.

When the practical initialization is used, the results differsignificantly. For white Gaussian noise, the MultSRP algorithmwith Wiener filtering and a warm start from MoSRP gives thebest performance. Wiener filtering generally appears to helpboth SRP methods, especially in the low-SNR regime. Themost accurate from among the SRP methods almost alwaysout-performs traditional SRP localization with the exceptionbeing in the case of crowd noise. This, of course, is mostlyan issue of initialization in the SRP methods. Once initializedwith the result of the sequential method, MultSRP with a warmstart performs significantly better.

Run times averaged over 50 trials using the practical ini-tialization are shown in Fig. 7. The MultSRP method is byfar the most efficient. This is mainly due to its simplicityand the fact that computations account for all the sources atonce. The MoSRP method varies linearly with the number of

Fig. 6: Localization errors for crowd noise and a practicalinitialization. The MultSRP model with a warm start performsbest in all cases.

Fig. 7: Average run times for noisy, 10-second speech mix-tures. All experiments were conducted on an iMac with 16GB of RAM and a 3.4 GHz processor.

sources because it computes source-specific gradients. We notethat these algorithms can be parallelized over sources (whereapplicable) and frequencies. We also note that the efficientform of the 2-source MultSRP gradient (see Appendix C)reduces its computation time by about a factor of 1.5.

Figure 8 shows the source separation results correspondingto the experiments of Figure 5. Separation performance wasmeasured via the Signal-to-Interference Ratio (SIR) metricwith the BSS Eval toolbox [40]. These results qualitativelymirror the localization results. We would expect this to bethe case since we are better able to isolate the energy cor-responding to each speaker with more accurate estimates oftheir DOAs.

B. Early Reflections

Figure 9 shows the result of a synthetic experiment withearly reflections. An 8-channel, square array of side-length 20

centimeters was positioned 10 centimeters from the middle ofa wall in a 3-dimensional room of size 5⇥5⇥5 meters. The im-age method [38] was applied to simulate reverberation for twospeakers using random sentences from the TSP database [37].In order to capture the effect of the early reflections, we setthe number of estimated sources to K = 4 (twice the truenumber). We can observe that the mixture of SRPs model(MoSRP) fails to find the source images while the multi-sourceSRP model (MultSRP) localizes them correctly. The MoSRPand MultSRP models achieve localization errors of e = 0.71

9

Fig. 8: Source separation results for 10-second speech mixturesin white Gaussian noise. The MultSRP method (with a warmstart) increasingly performs better than the others for moresources, especially for a low SNR.

−1 −0.5 0 0.5 1

4

4.5

5

5.5

6

Localization w/ Early Reflections

Source Positions

Source Images

MoSRP

MultSRP

Array

Wall

Fig. 9: Localization results with early reflections off of a wall.In this figure, the 3-dimensional room is viewed from above.

and e = 0.18 averaged over 10 trials with a similar setup (thesource positions were perturbed relative to the positions shownin Fig. 9). In this experiment, we used a simple initializationstrategy that places the initial DOA vectors near the top of thehemisphere (i.e. the center of Fig. 9).

C. Other SRP objective functionsWe applied the single- and multi-source optimization ap-

proaches described in Section V to the empirical output powerfunctions of the MVDR and data-independent and -dependentLCMV beamformers. However, we found that this does notperform nearly as well in practice as optimizing the proposedSRP likelihoods. This may be due to sensitivity to the inversionof Rf , which must be estimated from data, and to otherreal-world factors such as reverberation and overlap in theTF plane. In addition, the EM algorithm with a mixture ofMVDR objectives is quite slow since KF matrices bR(k)

f mustbe calculated and inverted in every iteration.

VIII. CONCLUSIONS

We used a probabilistic interpretation of the Steered Re-sponse Power (SRP) function to describe two algorithms for

the simultaneous source localization and separation problem.The first was based on a mixture-of-SRPs (MoSRP) modeland lead to a GEM algorithm for optimizing the sources’Directions-Of-Arrival (DOA). The second was based on amulti-source SRP (MultSRP) function that was robust tosource overlap in the Time-Frequency (TF) domain and leadto a simple gradient ascent scheme. Unlike in traditionalsequential methods, Wiener filtering was tied into the DOAoptimization procedure to make it robust to non-directionalinterference. Experiments showed that the proposed unifiedapproach outperforms a sequential one. We also showed thatthe MultSRP model can localize highly coherent sources withsignificant overlap in the TF plane, unlike the MoSRP model.

APPENDIX ADERIVATION OF SRP LIKELIHOOD

We derive a generalized form for the SRP likelihood ofSection IV-A. We assume a Gaussian distribution for each datavector xft (dependence on f and t are omitted for clarity):

L = �1

2

log

2⇡⌃

�� 1

2

(x� µ)

H⌃

�1(x� µ) , (47)

To find estimates bµ and b⌃ specific to x, we consider thefollowing generative model:

s ⇠ N (µs

,⌃

s

) , x|s ⇠ N (As,⌃

n

) . (48)

The general form of the posterior distribution [41] is:

s|x ⇠ N⇣

µs|x,⌃s|x

, (49)

µs|x = ⌃

s|x�

A

H⌃

�1n

x + ⌃

�1s

µs

, (50)

�1s|x = A

H⌃

�1n

A + ⌃

�1s

. (51)

Assuming a vague prior (i.e. ⌃

s

!1 I), we have:

bµ = E [x] = E [As + n] = Aµs|x = Bx , (52)

b

⌃ = E

h

(x� bµ) (x� bµ)

Hi

(53)

= E

As + n�Aµs|x

⌘⇣

As + n�Aµs|x

⌘H�

(54)

= E

A

s� µs|x

+ n

⌘⇣

A

s� µs|x

+ n

⌘H�

(55)

= A⌃

s|xAH

+ ⌃

n

(56)= (B + I)⌃

n

, (57)

where B = A

A

H⌃

�1n

A

⇤�1A

H⌃

�1n

. We substitute bµ andb

⌃ into (47), evaluate�

� b

� with Sylvester’s determinant theoremand b⌃

�1with the Woodbury identity, and simplify using the

fact that B is idempotent (i.e. B

2= B). Thus, we have:

L = �1

2

log

2⇡⌃

n

�� K

2

log 2� 1

2

x

HP

H⌃

�1n

Px , (58)

where P = I�B. Finally, if ⌃

n

= �

2I, we have:

L / � 1

2�

2x

HPx . (59)

10 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING

APPENDIX BGRADIENT OF SRP LIKELIHOOD

We derive the gradient of the multiple-source SRP likeli-hood with matrix calculus [42]. The DOA-dependent contri-bution to the SRP likelihood (27) from a single observed vectorx is:

L / 1

2�

2x

HBx , (60)

where B = A

A

HA

⇤�1A

H and:

Aiv =

1pM

e

j2⇡lf

c

Pu

mui⇥uv

. (61)

Applying the product rule, we identify three terms in thegradient corresponding to the gradients of:

P(1)= x

HAy , (62)

P(2)= z

H⇥

A

HA

⇤�1z , (63)

P(3)= y

HA

Hx , (64)

where y =

A

HA

⇤�1z and z = A

Hx are treated as

constants. Based on the result:

@ P(1)

@ ⇥uv= j

2⇡lf

c

X

i

muix⇤iAivyv , (65)

we can write:

@ P(1)

@ ⇥

= j

2⇡lf

c

m diag (x

⇤)A diag (y) , (66)

@ P(3)

@ ⇥

= �j 2⇡lf

c

m diag (x)A

⇤ diag (y

⇤) . (67)

With some abuse of notation, the second gradient term is:

@ P@ ⇥

(2)

= z

H

@

A

HA

⇤�1

@ ⇥

!

z (68)

= �z

H⇥

A

HA

⇤�1✓

@ A

HA

@ ⇥

A

HA

⇤�1z (69)

= �y

H

"

@ A

@ ⇥

◆H

A + A

H

@ A

@ ⇥

#

y (70)

= �

@ y

HA

Hq

@ ⇥

+

@ q

HAy

@ ⇥

(71)

= �j 2⇡lf

c

m

� diag (q)A

⇤ diag (y

⇤) · · ·

+ diag (q

⇤)A diag (y)

, (72)

where q = Ay is treated as a constant and we have applied theresults in (66)-(67) to (71). The temporary abuse of notationis tolerable once we observe that the non-zero terms in thetensor-valued gradient expressions form a matrix.

It follows that:

@ P@ ⇥

=

3X

i=1

@ P@ ⇥

(i)

(73)

=

4⇡lf

c

m

� Im [diag (x

⇤)A diag (y)] · · ·

+ Im [diag (q

⇤)A diag (y)]

(74)

=

4⇡lf

c

m

�Im⇥

A��

x

⇤y

>�⇤+ Im

A��

q

⇤y

>�⇤� (75)

=

4⇡lf

c

m

�Imh

A��

y x

H�>i

+ Imh

A��

y q

H�>i⌘

.

(76)

Aggregating over time and consolidating terms, we have:

@ P@ ⇥

=

4⇡lf

c

m Im⇥

A

⇤ ��

PC

H�⇤

, (77)

where C =

A

HA

⇤�1A

HR and P = I�B is the projection

matrix defined in (26). The final result for the gradient is:

@ L@ ⇥

=

1

2�

2

@ P@ ⇥

=

1

2

2⇡lf

c

m Im⇥

A

⇤ ��

PC

H�⇤

. (78)

APPENDIX CEFFICIENT COMPUTATION OF GRADIENT AND LIKELIHOOD

FOR 2 SOURCES

The gradient derived in Appendix B can be implementedmore efficiently in the case of K = 2 sources. We start byexplicitly evaluating the matrix inverse:

A

HA

⇤�1=

1

d

1 �a

H1 a2

�a

H2 a1 1

, (79)

where d = 1� |aH1 a2|2. Substituting this into (78), expanding,

and re-arranging, we have:

@ L@ ⇥

=

1

2

2⇡lf

c d

m Im⇥

z11 � v21 z12 � y , z22 � v12 z21 + y

,

(80)

where:

y =

r12v221 � (r11 + r22) v21 + r21

d

a

⇤1 � a2 , (81)

qi = Rai , (82)zij = a

⇤i � qj , (83)

vij = a

Hi aj , (84)

rij = a

Hi Raj = a

Hi qj . (85)

The likelihood can be computed efficiently as:

L = � 1

2�

2

�kXk2F +

r11 + r22 � 2 Re [v21r12]

d

. (86)

REFERENCES

[1] H. K. van Trees, Detection, Estimation, and Modulation Theory:Optimum Array Processing (Part IV), Wiley, 2002.

[2] S. Haykin, Adaptive Filter Theory, Prentice Hall, 3 edition, 1995.[3] J. Benesty, J. Chen, and Y. Huang, Topics in Signal Processing:

Microphone Array Signal Processing, vol. 1, Springer, 2008.[4] L. J. Griffiths and C. W. Jim, “An alternative approach to linearly

constrained adaptive beamforming,” IEEE Transactions on Antennasand Propagation, vol. 30, no. 1, pp. 27–34, 1982.

11

[5] Y. Mahieux C. Marro and K. U. Simmer, “Analysis of noise reductionand dereverberation techniques based on microphone arrays with post-filtering,” IEEE Transactions on Speech and Audio Processing, vol. 6,no. 3, 1998.

[6] J.-F. Cardoso and A. Souloumiac, “Blind beamforming for non gaussiansignals,” IEE Proceedings, vol. 140, pp. 362–370, 1993.

[7] S. Gannot and T. G. Dvorkind, “Microphone array speaker localizersusing spatio-temporal information,” EURASIP Journal on Advances inSignal Processing, vol. 2006, pp. 1–17, 2006.

[8] W. Herbordt, Sound Capture for Human-Machine Interfaces, Springer,2005.

[9] I. J. Tashev, Sound Capture and Processing: Practical Approaches,Wiley, 2009.

[10] S. Gazor, S. Affes, and Y. Grenier, “Wideband multi-source beam-forming with adaptive array location calibration and direction finding,”International Conference on Acoustics, Speech, and Signal Processing(ICASSP), vol. 3, pp. 1904–1907, 1995.

[11] J.-M. Valin, F. Michaud, and J. Rouat, “Robust localization and trackingof simultaneous moving sound sources using beamforming and particlefiltering,” Robot. Auton. Syst., vol. 55, no. 3, pp. 216–228, 2007.

[12] S. T. Birchfield and D. K. Gillmor, “Fast Bayesian acoustic localiza-tion,” IEEE International Conference on Acoustics, Speech, and SignalProcessing (ICASSP), vol. 2, pp. 1793–1796, 2002.

[13] C. H. Knapp and G. C. Carter, “The generalized correlation method forestimation of time delay,” IEEE Transactions on Acoustics, Speech, andSignal Processing, vol. 24, no. 4, pp. 320–327, 1976.

[14] S. T. Birchfield and D. K. Gillmor, “Acoustic source direction byhemisphere sampling,” IEEE Conference on Acoustics, Speech, andSignal Processing, vol. 5, pp. 3053–3056, 2001.

[15] R. O. Schmidt, “Multiple emitter location and signal parameter estima-tion,” IEEE Transactions on Antennas and Propagation, vol. 34, no. 3,pp. 276–280, 1986.

[16] R. Kumaresan and D. W. Tufts, “Estimating the angles of arrival ofmultiple plane waves,” IEEE Transactions on Aerospace and ElectronicSystems, vol. 19, no. 1, pp. 134–139, 1983.

[17] R. Roy and T. Kailath, “ESPRIT - estimation of signal parametersvia rotational invariance techniques,” IEEE Transactions on Acoustics,Speech and Signal Processing, vol. 37, no. 7, pp. 984–995, 1989.

[18] M. R. Azimi-Sadjadi, A. Pezeshki, and L. L. Scharf, “WidebandDOA estimation algorithms for multiple target detection and trackingusing unattended acoustic sensors,” in Proceedings of the Society ofPhotographic Instrumentation Engineers, 2004, vol. 5417, pp. 1–11.

[19] K. H. Knuth, “Bayesian source separation and localization,” Proceedingsof SPIE: Bayesian Inference for Inverse Problems, vol. 3459, pp. 147–158, 1998.

[20] B. Lee, Robust speech recognition in a car using a microphone array,Ph.D. thesis, University of Illinois at Urbana-Champaign, 2006.

[21] K. Harmanci, J. Tabrikian, and J.L. Krolik, “Relationships betweenadaptive minimum variance beamforming and optimal source localiza-tion,” IEEE Transactions on Signal Processing, vol. 48, no. 1, 2000.

[22] C. Zhang, D. Florencio, D. E. Ba, and Z. Zhang, “Maximum likelihoodsound source localization and beamforming for directional microphonearrays in distributed meetings,” IEEE Transactions on Multimedia, vol.10, no. 3, pp. 528–548, 2008.

[23] N. Q. K. Duong, E. Vincent, and R. Gribonval, “Under-determinedreverberant audio source separation using a full-rank spatial covariancemodel,” IEEE Transactions on Audio, Speech, and Language Processing,vol. 18, no. 7, pp. 1830–1840, 2010.

[24] J. Dmochowski, J. Benesty, and S. Affes, “Direction of arrival estimationusing the parameterized spatial correlation matrix,” IEEE Transactionson Audio, Speech, and Language Processing, vol. 15, no. 4, pp. 1327–1339, 2007.

[25] Y. Li and D Wang, “On the optimality of ideal binary time-frequencymasks,” Elsevier - Speech Communication, vol. 51, pp. 230–239, 2009.

[26] S. Rickard and O. Yilmaz, “On the approximate W-disjoint orthogo-nality of speech,” IEEE Conference on Acoustics, Speech, and SignalProcessing (CASSP), vol. 1, pp. 529–532, 2002.

[27] O. Yilmaz and S. Rickard, “Blind separation of speech mixtures viatime-frequency masking,” IEEE Transactions on Signal Processing, vol.52, pp. 1830–1847, 2004.

[28] S. Araki, H. Sawada, R. Mukai, and S. Makino, “Underdeterminedblind sparse source separation for arbitrarily arranged multiple sensors,”Signal Processing, vol. 87, no. 8, pp. 1833 – 1847, 2007.

[29] S. M. Naqvi, W. Wang, M. S. Khan, M. Barnard, and J. A. Chambers,“Multimodal (audiovisual) source separation exploiting multi-speakertracking, robust beamforming and time-frequency masking,” SignalProcessing, IET, vol. 6, no. 5, pp. 466–477, 2012.

[30] X. Zhong and J. R. Hopgood, “Time-frequency masking based multipleacoustic source tracking applying Rao-Blackwellized Monte Carlo dataassociation,” IEEE 15th Workshop on Statistical Signal Processing, pp.253–256, 2009.

[31] M. Wax and A. Leshem, “Joint estimation of time delays and directionsof arrival of multiple reflections of a known signal,” IEEE Transactionson Signal Processing, vol. 45, no. 10, pp. 2477–2484, 1997.

[32] M. L. Seltzer, “Bridging the gap: Towards a unified framework forhands-free speech recognition using microphone arrays,” IEEE Hands-Free Speech Communication and Microphone Arrays (HSCMA), pp.104–107, 2008.

[33] X. Zhao, J. Tang, L. Zhou, and Z. Wu, “A fast search method of steeredresponse power with small-aperture microphone array for sound sourcelocalization,” Journal of Electronics (China), vol. 30, no. 5, pp. 483–490, 2013.

[34] Hoang Do, H.F. Silverman, and Ying Yu, “A real-time srp-phat sourcelocation implementation using stochastic region contraction(src) on alarge-aperture microphone array,” in IEEE Internatinal Conference onAcoustics, Speech and Signal Processing (ICASSP), 2007, vol. 1.

[35] A. P. Dempster, N. M. Laird, and D. B. Rubin, “Maximum likelihoodfrom incomplete data via the em algorithm,” Journal of the RoyalStatistical Society, Series B, vol. 39, no. 1, pp. 1–38, 1977.

[36] R. M. Neal and G. E. Hinton, “A view of the em algorithm that justifiesincremental, sparse, and other variants,” in Learning in GraphicalModels. 1998, pp. 355 – 368, Kluwer Academic Publishers.

[37] P. Kabal, “TSP speech database,” 2002, Telecommunications and SignalProcessing Lab, McGill University.

[38] J. Allen and D. Berkley, “Image method for efficiently simulating small-room acoustics,” J. Acoust. Soc. Am., vol. 65, no. 4, pp. 943–950, 1979.

[39] J. Traa and P. Smaragdis, “Multichannel source separation and trackingwith ransac and directional statistics,” IEEE/ACM Transactions onAudio, Speech, and Language Processing (TASLP), vol. 22, pp. 2233–2243, 2014.

[40] E. Vincent, R. Gribonval, and C. Fevotte, “Performance measurementin blind audio source separation,” IEEE Transactions on Audio, Speech,and Language Processing, vol. 14, no. 4, pp. 1462 –1469, 2006.

[41] K. P. Murphy, Machine Learning: A Probabilistic Perspective, MITPress, 2012.

[42] K. B. Petersen and M. S. Pedersen, “The matrix cookbook,” 2012.

Johannes Traa ([email protected]) received theB.S. degree in electrical engineering from North-western University, Evanston, IL, in 2011 and theM.S. degree in electrical and computer engineeringfrom the University of Illinois at Urbana-Champaign(UIUC), Urbana, IL, in 2013. He is currently pur-suing the Ph.D. degree in electrical and computerengineering at UIUC.

He has been a Research Assistant with ParisSmaragdis at UIUC since 2011. In the summers of2012-2014, he interned with the theory group at

Lyric Labs, Analog Devices in Boston, MA. His research interests includeaudio source separation and localization, sound mixture analysis with additivemodeling techniques like non-negative matrix factorization, and applicationsof various areas of statistics (e.g. compositional, directional) to audio prob-lems.

12 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING

David Wingate ([email protected]) received hisBS and MS degrees in computer science fromBrigham Young University in 2002 and 2004, anda PhD in computer science from the University ofMichigan in 2008. He was a postdoctoral fellow atMIT from 2008-2010 with a joint appointment in theComputer Science and Artificial Intelligence Labora-tory and the Computational Cognitive Science groupin the Brain and Cognitive Science Department.From 2010-2012 he was a research scientist at MITwith a joint appointment in BCS and the Laboratory

for Information and Decision Systems. From 2012-2015 he was a researchscientist at Analog Devices, Inc. in their machine learning group, where heworked on hardware accelerated Bayesian inference and audio processing.Since 2015 he is an assistant professor at Brigham Young University. Hisresearch interests include probabilistic programming, reinforcement learning,deep learning, and the intersection of hardware and machine learning.

Noah D. Stein ([email protected]) is a seniorresearch scientist at the Lyric Labs research unit ofAnalog Devices, Inc in Cambridge, MA. He receivedthe B.S. degree in electrical and computer engineer-ing from Cornell University in 2005 and the Ph.D.in electrical engineering from MITs Laboratory forInformation and Decision Systems in 2011.

Noahs research at Lyric Labs focuses on algo-rithms for audio source separation with an empha-sis on practical machine learning approaches forexploiting directional information as alternatives to

beamforming. His doctoral research applied semidefinite programming andother algebraic methods in optimization to theoretical questions in gametheory.

Paris Smaragdis ([email protected]) is an assistantprofessor in the Computer Science Department andthe Electrical and Computer Science Department atthe University of Illinois at Urbana-Champaign, aswell as a senior research scientist at Adobe Research.Prior to that he was a research scientist at MitsubishiElectric Research Labs, during which time he wasselected by MIT Technology Review as one ofthe top 35 young innovators of 2006. His researchinterests lie in the intersection of machine learningand signal processing. He is a Senior Member of the

IEEE.