Tampere University of Technology, Media Technologies...

5
SELF-LOCALIZATION OF WIRELESS ACOUSTIC SENSORS IN MEETING ROOMS Mikko Parviainen, Pasi Pertilä * Tampere University of Technology, Department of Signal Processing, Tampere, Finland {mikko.p.parviainen, pasi.pertila}@tut.fi Matti S. Hämäläinen Media Technologies Laboratory Nokia Research Center Tampere, Finland [email protected] ABSTRACT This paper presents a passive acoustic self-localization and synchro- nization system, which estimates the positions of wireless acoustic sensors utilizing the signals emitted by the persons present in the same room. The system is designed to utilize common off-the-shelf devices such as mobile phones. Once devices are self-localized and synchronized, the system could be utilized by traditional array pro- cessing methods. The proposed calibration system is evaluated with real recordings from meeting scenarios. The proposed system builds on earlier work with the added contribution of this work is i) increas- ing the accuracy of positioning, and ii) introduction data-driven data association. The results show that improvement over the existing methods in all tested recordings with 10 smartphones. 1. INTRODUCTION Traditional microphone array methods such as beamforming and source localization [1] require that microphone positions are known and that there is no temporal offsets between the captured signals, i.e., the signals are synchronized. The advancement of modern com- munication devices such as smartphones, tablets, and more recently wearable devices have created ubiquitous microphone arrays. Un- fortunately, such microphone locations and their temporal offsets are generally not available in accurate enough form, which would allow the direct utilization of the traditional array processing methods. The problem of simultaneously locating devices, estimating the temporal offsets, and locating external sources using only passive lis- tening, can be principally solved by minimizing a global cost func- tion that incorporates all the unknowns and the corresponding mea- surements [3]. However, this approach requires a good initial guess to avoid converging to local minima. In [5] a self-localization so- lution that estimates distances between devices from diffuse sound field is proposed. The positions of microphones are estimated using Multidimensional Scaling (MDS) [5]. Recent advances in self-localization [6] and temporal offset esti- mation [7] provide accurate initial guesses when the assumptions of the methods are met. Once the initial positions and temporal offsets are available, traditional source localization techniques [1] can be used to obtain initial locations. Since an error in sensor location can in some cases lead to double the error in source localization [8], the estimates of microphone positions and offsets should be further re- fined. Fortunately, the minimization of the global cost-function can be performed, once the captured data is assigned to its correspond- ing source. The assignment in itself is a separate research problem * This work is funded by Nokia Research Center and Finnish Academy project no. 138803. referred as data-association, which also can deal with detecting mea- surement errors caused by clutter and noise (see e.g. [9, Ch. 16]). In this work, we propose a "divide and conquer" approach for the problem of microphone self-localization and in meeting room sce- nario, where devices are static on a table, and the speakers are seated at the table. The novelty of this work is the combination of the initial guess methods with a provided data-association technique. The per- formance of self-localization using actual smartphone recordings is contrasted to the initial estimates to demonstrate accuracy improve- ment. 2. FORMULATION Let mi R 3 be the ith receiver position i 1,...,N , with N microphones. In an anechoic room the signal mi (t) can be modeled as a delayed source signal s k (t) as mi (t)= sk(t - τ k i )+ ni (t), (1) where t is time, k [1,...,K] denotes source index with K sources, and τ k i is time of arrival (TOA) from source k to the ith microphone τ k i = c -1 s k - mi + δi , (2) where δi is unknown time offset, c is speed of sound, and s k , mi R 3 are source and microphone positions. The time difference of arrival (TDOA) between microphone pair {i, j } for source k is τ k i,j τ k i - τ k j = c -1 (sk - mi -sk - mj )+ δij , (3) where pairwise time offset is δij δi - δj . The vector of all TDOA values is denoted τ =[τ 1 1,2 , τ 1 1,3 ,..., τ K N-1,N ] T , and τ R KP . TDOA estimates can be obtained e.g. using correlation ([10]) be- tween all P = N (N - 1)/2 unique microphone pairs for each source. The General Self-Localization Problem (GSLP) is solved by the following minimization problem [3] J ( ˆ S, ˆ M, ˆ δ)= (4) argmin S,M,δ {i,j,k} c -1 (sk - mi -sk - mj )+ δij - τi,j,k 2 , where the sum is over all K sources and P microphone pairs, S = [s1, s2,..., sK] T , M =[m1, m2,..., mN ] T , δ =[δ1, δ2,..., δN ] T . Note that the result of the minimization is subject to arbitrary ro- tation, reflection, translation, and arbitrary common time offset. This leads to Nu = 3(N + K)+ N - 7 number of unknown variables, which can not exceed the Nm = S(N - 1) independent measurements [3]. The degrees of freedom (DOF) is defined here as © 2014 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.

Transcript of Tampere University of Technology, Media Technologies...

SELF-LOCALIZATION OF WIRELESS ACOUSTIC SENSORS IN MEETING ROOMS

Mikko Parviainen, Pasi Pertilä!

Tampere University of Technology,Department of Signal Processing, Tampere, Finland

{mikko.p.parviainen, pasi.pertila}@tut.fi

Matti S. Hämäläinen

Media Technologies LaboratoryNokia Research Center

Tampere, [email protected]

ABSTRACT

This paper presents a passive acoustic self-localization and synchro-nization system, which estimates the positions of wireless acousticsensors utilizing the signals emitted by the persons present in thesame room. The system is designed to utilize common off-the-shelfdevices such as mobile phones. Once devices are self-localized andsynchronized, the system could be utilized by traditional array pro-cessing methods. The proposed calibration system is evaluated withreal recordings from meeting scenarios. The proposed system buildson earlier work with the added contribution of this work is i) increas-ing the accuracy of positioning, and ii) introduction data-driven dataassociation. The results show that improvement over the existingmethods in all tested recordings with 10 smartphones.

1. INTRODUCTION

Traditional microphone array methods such as beamforming andsource localization [1] require that microphone positions are knownand that there is no temporal offsets between the captured signals,i.e., the signals are synchronized. The advancement of modern com-munication devices such as smartphones, tablets, and more recentlywearable devices have created ubiquitous microphone arrays. Un-fortunately, such microphone locations and their temporal offsets aregenerally not available in accurate enough form, which would allowthe direct utilization of the traditional array processing methods.

The problem of simultaneously locating devices, estimating thetemporal offsets, and locating external sources using only passive lis-tening, can be principally solved by minimizing a global cost func-tion that incorporates all the unknowns and the corresponding mea-surements [3]. However, this approach requires a good initial guessto avoid converging to local minima. In [5] a self-localization so-lution that estimates distances between devices from diffuse soundfield is proposed. The positions of microphones are estimated usingMultidimensional Scaling (MDS) [5].

Recent advances in self-localization [6] and temporal offset esti-mation [7] provide accurate initial guesses when the assumptions ofthe methods are met. Once the initial positions and temporal offsetsare available, traditional source localization techniques [1] can beused to obtain initial locations. Since an error in sensor location canin some cases lead to double the error in source localization [8], theestimates of microphone positions and offsets should be further re-fined. Fortunately, the minimization of the global cost-function canbe performed, once the captured data is assigned to its correspond-ing source. The assignment in itself is a separate research problem

!This work is funded by Nokia Research Center and Finnish Academyproject no. 138803.

referred as data-association, which also can deal with detecting mea-surement errors caused by clutter and noise (see e.g. [9, Ch. 16]).

In this work, we propose a "divide and conquer" approach for theproblem of microphone self-localization and in meeting room sce-nario, where devices are static on a table, and the speakers are seatedat the table. The novelty of this work is the combination of the initialguess methods with a provided data-association technique. The per-formance of self-localization using actual smartphone recordings iscontrasted to the initial estimates to demonstrate accuracy improve-ment.

2. FORMULATION

Let mi ! R3 be the ith receiver position i ! 1, . . . , N , with Nmicrophones. In an anechoic room the signal mi(t) can be modeledas a delayed source signal sk(t) as

mi(t) = sk(t" !ki ) + ni(t), (1)

where t is time, k ! [1, . . . ,K] denotes source index with Ksources, and !k

i is time of arrival (TOA) from source k to the ithmicrophone

!ki = c"1#sk "mi#+ "i, (2)

where "i is unknown time offset, c is speed of sound, and sk,mi !R3 are source and microphone positions. The time difference ofarrival (TDOA) between microphone pair {i, j} for source k is

!ki,j ! !k

i " !kj = c"1(#sk "mi# " #sk "mj#) + "ij , (3)

where pairwise time offset is "ij ! "i " "j . The vector of all TDOAvalues is denoted ! = [!1

1,2, !11,3, . . . , !

KN"1,N ]T , and ! ! RKP .

TDOA estimates can be obtained e.g. using correlation ([10]) be-tween all P = N(N " 1)/2 unique microphone pairs for eachsource. The General Self-Localization Problem (GSLP) is solvedby the following minimization problem [3]

J(S, M, ") = (4)

argminS,M,"

!

#{i,j,k}

"c"1(#sk "mi# " #sk "mj#) + "ij " !i,j,k

#2,

where the sum is over all K sources and P microphone pairs, S =[s1, s2, . . . , sK ]T , M = [m1,m2, . . . ,mN ]T, " = ["1, "2, . . . , "N ]T.Note that the result of the minimization is subject to arbitrary ro-tation, reflection, translation, and arbitrary common time offset.This leads to Nu = 3(N + K) + N " 7 number of unknownvariables, which can not exceed the Nm = S(N " 1) independentmeasurements [3]. The degrees of freedom (DOF) is defined here as

© 2014 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.

Fig. 1: System diagram.

DOF = Nm"Nu, where DOF $ 0 should be forced. The redundantDOFs are removed by fixing a coordinate system m1 = [0, 0, 0]T

(translation), m2 = [m2x, 0, 0]T and m3 = [m3x,m3y, 0]

T. Fur-thermore, the clock offset of the first microphone is set to zero i.e."1 = 0.

3. THE SELF-LOCALIZATION SYSTEM

The proposed system (Figure 1) incorporates previously developedtechniques for processing speech signal in meeting room scenario.These include TDOA estimation (A), temporal offset estimation (B),initial microphone position estimation (C), and sound source local-ization (D). These techniques combined with the proposed data as-sociation scheme (E) enable the iterative self-localization (F). Thefundamental goal is to enhance initial microphone position and off-set estimates provided by (B), and (C).

3.1. Time Difference of Arrival (TDOA) Estimation (A)

Time-delay estimation is one possible key to self-localization, sinceit is a function of all the unknown variables (2), and it can be di-rectly measured from the sensors. TDOA estimates are utilized by allthe subsystems, which is computationally efficient. The peak indexvalue of the generalized cross-correlation with a weighting functionis used to calculate the TDOA !ij between the signals of microphonepair i, j, refer to [10]. The PHAT weight function removes the am-plitude information and in practice shapes the correlation functionto make the peak of cross correlation function more prominent androbustness [11].

The TDOA estimates contain a significant amount of instanceswhich do not originate from a speech source. Therefore a sequentialfilter is applied as follows. If a TDOA estimate in the current framehas changed over a threshold compared to the previous frame in anymicrophone pair, then the current frame is labeled as an outlier andis not processed by the system.

Fig. 2: Wireless acoustic sensors placed on a the table of a meetingroom. The arrows indicate sound wave propagation directions whensound sources are emitting at endfire positions making estimation of!maxij and !min

ij possible.

3.2. Offset Estimation (B)

The proposed method is designed for an ad hoc device network. Toutilize the data collected by devices, a common time base needs tobe established. This is done by offset estimation subsystem. Theoffset estimation is conducted using the method presented in [7].The pairwise offsets "ij are obtained as follows.

"ij =12(!max

ij + !minij ), (5)

where !minij and !max

ij are the minimum and the maximum TDOAvalues for microphone pair (i, j). The proof of (5) is presented in [7].The observation of the minimum and maximum time delays requiresthat during the recording signals are emitted from the line that con-nects each microphone pair (i, j) (see Figure 2). Due to one missingdegrees of freedom, the method [7] produces microphone offset esti-mates "i, i = 2, . . . , N from the pairwise measurements (5), wherethe offset values are relative to the first microphone, which can be(arbitrarily) set to zero "1 = 0.

3.3. Initial Microphone Position Estimation (C)

The purpose of the initial microphone position is to roughly estimateminimum and the maximum TDOA values for microphone pairs.The method presented in [6] is used. The fundamental idea is to isto estimate the pairwise distance using speed of sound c

dij =c2(!max

ij " !minij ). (6)

From the distance matrix, consisting of all pairwise distances, thepositions of the microphones in a relative coordinate system are ob-tained using MDS [12].

3.4. Source Localization (D)

Closed-form source localization techniques such as [13] and [14] areattracting due to computational efficiency, but involve a lineariza-tion of quadratic equations. The accuracy of closed-form may besufficient enough for providing initialization for source variables ingeneral self-localization problem (4) [8]. In testing the closed-formsource localization methods presented [13] and [14], they turn out

to be suffering from inaccuracies in microphone positioning and off-set estimation provided by subsystem (C) with the data used in thiswork, which is a similar finding to [8].

Source localization is thus done via iterative optimizationby solving (4) for each source individually. Microphone posi-tions are considered known in (4). Microphone position estimatesm1, . . . , mN are obtained from subsystem (B) and pairwise tempo-ral offset estimates "12, . . . , "N"1,N from subsystem (C). Pairwiseoffset values are subtracted from TDOA values to enable traditionalsource localization methods that assume perfectly synchronizedmicrophones with known locations.

In optimization of the cost function (4) Matlab function lsqnonlin[15] is used with trust-region reflective algorithm. The number of it-erations allowed is set to 5000. Termination tolerances for variableschange and objective function change are set to to 1 · 10"12.

3.5. Data Association (E)

The purpose of data association is to determine to which soundsource TDOA measurements belong. The data association (DA)scheme in the proposed system is based on assumption that sourcesare non-moving and only one source is active at a time. However,this kind of assumption is reasonable for instance in a meeting, inwhich it is common that people are sitting at a table and talkingone at a time. A more advanced DA technique is required if theconditions are not met [9, Ch. 16]. The fundamental idea is to detectfrom TDOA measurements changes that result from the fact thatactive sources has changed.

There are a total of P = N(N"1)/2 TDOA measurements pertime frame. Thus the input signal ! to the DA subsystem is a P %Tmatrix, where T is amount of frames used for TDOA estimation.Since multiple source positions can be mapped into the same TDOAvalue, all P pairs of TDOA values should be considered when try-ing to identify a sound source. We propose to use Principal Com-ponent Analysis (PCA) to obtain a reduced set of TDOA features.PCA seeks a projection in least squares sense that best presents thedata to enable simple detection methods such as clustering or peak-picking [16].

PCA results in principal components P and S scores. The orig-inal data ! is reconstructed as follows.

! = SPT (7)

The operation of DA subsystem is as follows. (I) Perform PCAfor the data. (II) Estimate probability density of the reduced data(i.e. all data up to current frame). (III) Detect peaks from densityestimate. (IV) Find time indices that correspond to the density max-ima. (V) Use the indices to find appropriate TDOA values.

Figure 3 illustrates the input data, reduced data and its use foridentifying sources. Figure 3a presents input data ! i.e. the original45 dimensional TDOA data over a recording. Each dimension ispresented by a different color. Figure 3b presents the data mappedfrom 45 to 2 over the recording. The blue line corresponds to the firstprincipal component and the green line corresponds to the secondprincepal component. Approximately constant segments e.g. fromframe 125 to 175 represent time periods when a particular source isactive. Transitions to another static segment corresponds to sourcechange. Figure 3c presents probability density estimation of reduceddata corresponding to the first two PCAs, where peaks correspondto sound sources. Comparing the panels 3c and 3b one can observethat peaks occur in probability density estimate at the same locationswhere there are a lot of data points in Figure 3b.

(a) (b)

(c) (d)

Fig. 3: Illustration of dimensionality reduction. Panel a is input datato system i.e. all TDOA pairs of a recording. Panel b is data mappedto a lower dimensional using PCA. The blue line corresponds tothe first principal component and the green line corresponds to thesecond principal component. Panel c presents probability densityestimate of the reduced data. Panel d is the resulting sound sourcelabeling over a recording.

By detecting peaks from probability density estimate Figure 3cand searching for corresponding time instants from TDOA, one canlabel frames by a unique source ID number. The resulting TDOAlabeling using the proposed DA is illustrated in Figure 3d.

3.5.1. Iterative Self-localization (F)

The iterative self-localization solves (4) via optimization. In theused data the number of microphones is N = 10 and the numberof sources is K = 4. Taking into account the redundant degreesof freedom, the number of unknown variables is 45. Using the in-formation provided by subsystems (B)-(E), it is possible to initializethe optimization problem (4) with good values and thus increase theprobability of convergence to a reasonable solution.

The subsystem (F) outputs microphone position estimatesm1, . . . , mN as well as source position estimates s1, . . . , sK ofwhich the former is in the primary interest of this system.

4. DATA DESCRIPTION

Data is recorded using ten Nokia N900 mobile handsets runningMaemo operating system. Each device records the data using itsmicrophone with an added Sennheiser MKE 2P-C microphone at-tached near the microphone inlet for reference purposes. Samplingrate is 48000 Hz and bit depth is 16 bits. The analysis window inTDOA estimation (A) is 8192 samples (approximately 171 ms) with50 % overlap.

(a) Setup in Recording ID 1 (b) Setup in Recording ID 3

Fig. 4: Recording room 1

The recordings were made in meeting rooms. The scenariosimitate a meeting of four participants. The participants are seatedaround a table and N900s are placed on the table in front of meet-ing participants. In the actual recording, each participant utterancesa sentence after which another participants utterances another sen-tence etc. The length of each recording is approximately 1 minute30 s. In all recordings except Recording ID 3 participants are placedapproximately at the same positions with respect to devices. How-ever, they are slightly altering their posture from one recording toanother. Figure 4a illustrates the scenario. In Recording ID 3 oneparticipant is sitting at the end of the table as illustrated in Figure 4b.Position of devices in each room are same.

The reference microphone positions for evaluation were ob-tained with a tape measure.

5. RESULTS

The presentation of results focuses on positioning accuracy. Westress that the proposed data association method is novel and hasan essential role in the operation of the system.

The results obtained with the proposed system are presented aserror to the ground truth coordinates. The error is defined as

e =1N

N!

i=1

$%%& 1D

D!

d=1

(mi "mi)2, (8)

where N is the number of microphones, D is the dimension (hereD = 3)1 , location mi is the location estimate and mi is the groundtruth position of microphone i. Table 1 presents the results obtainedin six real recordings. The einit is the error after initial micro-phone positioning (B) and eenhanced is the error after iterative self-localization (F). The coordinate estimates are rotated, translated andreflected to match before analyzing the distance in (8).

Comparing the error using paired t-test between initial micro-phone positioning (B) and iterative self-localization, it can be seenthat the results were improved in a statististically significant way(p<0.05).

6. CONCLUSIONS

A self-localization system utilizing an ad hoc device network waspresented. The system is able to estimate positions of devices from

1Six sources should be used in conjunction with 10 microphones to resultin an unambiguous estimate. The recordings contain only four sound sources.Still, the optimization of (4) converges to a solution near to ground truthmicrophone positions.

Table 1: The performance of the system in meeting room recordings.einit is the error between the ground truth and the self-localizationsystem [6]. eenhanced is the error obtained using the proposed sys-tem.

Recording ID [m]einit eenchanced

1 0.167 0.1312 0.142 0.1283 0.112 0.0724 0.161 0.1475 0.132 0.1276 0.129 0.123

acoustic measurement. For instance, in a meeting scenario, the par-ticipants can use their mobile devices to establish a network. Usingthe proposed system the network can be self-localized and thus beused e.g. speaker localization, and annotation. All of the mentionedapplications of the can be used for enhanced teleconferencing expe-rience. The system was tested with data recorded in two meetingrooms. The recordings imitate a meeting scenario; participants areseated at a table and one person is speaking at a time. The proposedsystem achieves root-mean-square device position error of 7 – 15cm.

7. REFERENCES

[1] J.C. Chen, Kung Yao, and R.E. Hudson, “Source localizationand beamforming,” IEEE Signal Process. Mag., vol. 19, no. 2,pp. 30 – 39, 2002.

[2] Sebastian Thrun, “Affine structure from sound,” in In NIPS.2005, pp. 1353–1360, MIT Press.

[3] N. Ono, H. Kohno, N. Ito, and S. Sagayama, “Blind align-ment of asynchronously recorded signals for distributed mi-crophone array,” in Applications of Signal Processing to Audioand Acoustics, 2009. WASPAA ’09. IEEE Workshop on, 2009,pp. 161–164.

[4] V.C. Raykar, I.V. Kozintsev, and R. Lienhart, “Position calibra-tion of microphones and loudspeakers in distributed computingplatforms,” Speech and Audio Processing, IEEE Transactionson, vol. 13, no. 1, pp. 70 – 83, jan. 2005.

[5] Iain McCowan, Mike Lincoln, and Ivan Himawan, “Micro-phone array shape calibration in diffuse noise fields,” IEEETransactions on Audio, Speech, and Language Processing, vol.16, no. 3, pp. 666–670, 2008.

[6] Pasi Pertilä, Mikael Mieskolainen, and Matti Hämäläinen,“Passive Self-Localization of microphones using ambientsounds,” in 20th European Signal Processing Conference 2012(EUSIPCO 2012), Bucharest, Romania, Aug 2012.

[7] P. Pertilä, M.S. Hämäläinen, and M. Mieskolainen, “Passivetemporal offset estimation of multichannel recordings of an ad-hoc microphone array,” IEEE Transactions on Audio, Speech,and Language Processing, vol. 21, no. 11, pp. 2393–2402,2013.

[8] Wouter van Herpen, Sriram Srinivasan, and Piet Sommen, “Er-ror Analysis on Source Localization in Ad-Hoc Wireless Mi-crophone Networks,” in International Workshop on AcousticEcho and Noise Control (IWAENC), 2010.

[9] S. Haykin and K.J.R. Liu, Handbook on Array Processing andSensor Networks, Adaptive and Learning Systems for SignalProcessing, Communications and Control Series. Wiley, 2010.

[10] C. Knapp and G. Carter, “The generalized correlation methodfor estimation of time delay,” Acoustics, Speech and SignalProcessing, IEEE Transactions on, vol. 24, no. 4, pp. 320 –327, aug 1976.

[11] Bert Van Den Broeck, Alexander Bertrand, and Peter Karsmak-ers, “Time-domain generalized cross correlation phase trans-form sound source localization for small microphone arrays,”2012.

[12] I. Borg and P. Groenen, Modern Multidimensional Scaling,Springer-Verlag New York, Inc., 1997.

[13] J.O. Smith and J.S. Abel, “Closed-form least-squares sourcelocation estimation from range-difference measurements,”Acoustics, Speech and Signal Processing, IEEE Transactionson, vol. 35, no. 12, pp. 1661–1669, 1987.

[14] M.D. Gillette and H.F. Silverman, “A linear closed-form algo-rithm for source localization from time-differences of arrival,”Signal Processing Letters, IEEE, vol. 15, pp. 1–4, 2008.

[15] “Matlab function lsqnonlin,” referred on Jan 29, 2014.

[16] Richard O. Duda, Peter E. Hart, and David G. Stork, PatternClassification (2nd Edition), Wiley-Interscience, 2001.