Documente Academic
Documente Profesional
Documente Cultură
RESEARCH
Open Access
Abstract
This paper investigates real-time N-dimensional wideband sound source localization in outdoor (far-field) and lowdegree reverberation cases, using a simple N-microphone arrangement. Outdoor sound source localization in different
climates needs highly sensitive and high-performance microphones, which are very expensive. Reduction of the
microphone count is our goal. Time delay estimation (TDE)-based methods are common for N-dimensional wideband
sound source localization in outdoor cases using at least N + 1 microphones. These methods need numerical analysis
to solve closed-form non-linear equations leading to large computational overheads and a good initial guess to avoid
local minima. Combined TDE and intensity level difference or interaural level difference (ILD) methods can reduce
microphone counts to two for indoor two-dimensional cases. However, ILD-based methods need only one dominant
source for accurate localization. Also, using a linear array, two mirror points are produced simultaneously (half-plane
localization). We apply this method to outdoor cases and propose a novel approach for N-dimensional entire-space
outdoor far-field and low reverberation localization of a dominant wideband sound source using TDE, ILD, and headrelated transfer function (HRTF) simultaneously and only N microphones. Our proposed TDE-ILD-HRTF method tries to
solve the mentioned problems using source counting, noise reduction using spectral subtraction, and HRTF. A special
reflector is designed to avoid mirror points and source counting used to make sure that only one dominant source is
active in the localization area. The simple microphone arrangement used leads to linearization of the non-linear closedform equations as well as no need for initial guess. Experimental results indicate that our implemented method features
less than 0.2 degree error for angle of arrival and less than 10% error for three-dimensional location finding as well as
less than 150-ms processing time for localization of a typical wideband sound source such as a flying object (helicopter).
Keywords: Sound source localization; ITD; TDE; TDOA; PHAT; ILD; HRTF
1. Introduction
Source localization has been one of the fundamental
problems in sonar [1], radar [2], teleconferencing or
videoconferencing [3], mobile phone location [4], navigation and global positioning systems (GPS) [5], localization
of earthquake epicenters and underground explosions [6],
microphone arrays [7], robots [8], microseismic events in
mines [9], sensor networks [10,11], tactile interaction in
novel tangible human-computer interfaces [12], speaker
tracking [13], surveillance [14], and sound source tracking
[15]. Our goal is real-time sound source localization in
outdoor environments, which necessitates a few points to
be considered. For localizing such sound sources, a farfield assumption is usual. Furthermore, our experiments
* Correspondence: pourmohammad@aut.ac.ir
Electrical Engineering Department, Amirkabir University of Technology,
424 Hafez Ave., Tehran 15914, Iran
2013 Pourmohammad and Ahadi; licensee Springer. This is an open access article distributed under the terms of the Creative
Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
2. Literature review
Passive sound source localization methods, in general,
can be divided into direction of arrival (DOA) [16], time
difference of arrival (TDOA) or TDE or interaural time
difference (ITD)- [17-20], ILD- [21-24], and HRTFbased methods [25-30]. DOA-based beamforming and
sub-space methods typically need a large number of microphones for estimation of narrowband source locations in far-field cases and wideband source locations in
near-field cases. Also, they have higher processing needs
in comparison to other methods. Many localization
methods for near-field cases have been proposed in the
literature, such as maximum likelihood (ML), covariance
approximation, multiple signal estimation (MUSIC), and
estimation of signal parameters via rotational invariance
techniques (ESPRIT) [16]. However, these methods are
not applicable to the localization of wideband signal
sources in far-field cases with small number of microphones. On the other hand, ILD-based methods are
mostly applicable to the case of a single dominant sound
source (high signal-to-noise ratio (SNR)) [21-24]. TDEbased methods with high sampling rates are commonly
used for 2D and 3D high-accuracy wideband near-field
and far-field sound source localization. In the case of
ILD or TDE-based methods, minimum number of microphones required is three for 2D positioning and four
for the 3D case [17-20,31-39]. Finally, HRTF-based
methods are applicable only to the case of calculating arrival angle in azimuth or elevation [30].
TDE- and ILD-based outdoor far-field accurate wideband sound source localization in different climates
needs highly sensitive and high-performance microphones which are very expensive. In the last decade,
some papers were published which introduce 2D sound
source localization methods using just two microphones
in indoor cases using TDE- and ILD-based methods
simultaneously. However, it is noticeable that using ILDbased methods requires only one dominant source to be
active, and it is known that, by using a linear array in
the proposed TDE-ILD-based method, two mirror points
will be produced simultaneously (half-plane localization)
[40]. In this paper, we apply this method in outdoor (lowdegree reverberation) cases for a dominant sound source.
Page 2 of 19
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
Page 3 of 19
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
Y c k
X c k
In detail:
jH c k j
jY c k j
jX c k j
Yck
:
C c k X c k
3. Basic methods
3.1. HRTF
Page 4 of 19
s1 t
stT 1
n1 t
d1
s2 t
stT 2
n2 t ;
d2
and
st
n1 t
d1
s2 t
st
n2 t :
d2
and
E 1 wo s21 t dt
1 w 2
s t dt wo n21 t dt
d 21 o
10
H c k jH c k je jargH c k :
E 1 wo s22 t dt
1 w 2
s t dt wo n22 t dt:
d 22 o
11
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
According to (10) and (11), the received energy decreases in relation to the inverse of the square of the distance to the source. These equations lead us to a simple
relationship between the energies and distances:
E 1 d 21 E 2 d 22 ;
12
where wo d 21 n21 t d 22 n22 t dt is the error term. If
(x1,y1) is the coordinates of the first microphone, (x2,y2)
is the coordinates of the second microphone and (xs,ys)
is the coordinates of the sound source with respect to
the origin located at array center, then:
q
d 1 x1 xs 2 y1 ys 2
13
q
14
d 2 x2 xs 2 y2 ys 2 :
Now using (12), (13), and (14), we can localize the
sound source (Section 4.1.).
3.3. TDE-based localization
15
j2f
df :
S 1 f S 2 f e
17
1
;
jGs1 s2 f j
19
20
21
22
where
r s1 s2 E s1 t s2 t :
23
16
Page 5 of 19
18
25
26
27
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
Page 6 of 19
Figure 1 Hyperbolic geometrical location of 2D sound source localization using two microphones.
Assuming a single frequency sound source with a wavelength equal to to have a distance from the center of
two microphones equal to r, this source will be in farfield if [51-65]:
2D2
r>
;
31
32
33
34
Correct measurement of the time delay needs the distance between the two microphones to be:
D ;
35
2
since when D is greater than 2 , TD is greater than , and
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
therefore, time delay is measured as = (TD ). According to Figure 1, the cosine of the angle of arrival is
cos
:
D
D
D
D
36
38
This is an equation with two unknowns, x and y. Assuming the distances of both microphones from the origin to be R (D = 2R) and both located on x axis,
q q
x R2 y2 xR2 y2
21
:
39
vsound
Simplifying the above equation will result in:
8
y2 a:x2 b
>
>
>
>
4R2
<
a
1 ;
vsound 21 2
>
2
>
>
v
2
>
: b sound 21 R2
4
p2
21 vsound m
p
xs x2 ys y2
r 22
1 m
2
37
Page 7 of 19
43
1 2 2
k 1 r 1 R2s
2
44
x2 xs y 2 y s
1 2 2
k r R2s ;
2 2 s
45
and
where
k 21 x21 y21 ; k 22 x22 y22 and R2s x2s y2s :
40
yy
>
2
3 xx1 yy1
>
: 31
vsound
41
It is noticeable that these are non-linear equations
(Hyperbolic-intersection Closed-Form method) and numerical analysis should be used to calculate x and y,
which will increase localization processing times. Also in
this case, the solution may not converge.
46
R
s 1
ys
x2 y 2
k 2 2 r 2 2
2
and
xs
ys
x y
1 1
x2 y 2
1 2
1
k 1 r 21
2 1
Rs
:
1
k 22 r 22
2
48
If we define
1 x1 y1 1 1
p1
p
p2
1
2 x2 y2
49
and
q
1 x1 y1 1 k 21 r 12
q1
;
q2
k 22 r 22
2 x2 y 2
50
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
xs
ys
q1 p1 R2s
:
q2 p2 R2s
51
01 02
;
03
52
where
01 1p1 q1 p2 q2;
q
02 1p1 q1 p2 q2 2 p21 p22 q21 q22 ;
03 p21 p22 :
The positive root gives the square of distance from
source to origin. Substituting Rs into (51), the final
source coordinate will be obtained [40].
However, a rational solution requires prior information
of evaluation regions. It is known to us that, by using a
linear array, two mirror points will be produced simultaneously. Assuming two microphones are on x-axis
(y1 = y2 = 0) and have distance R from origin (Figure 1),
According to (49) and (50), we cannot find p and q.
Therefore, we cannot consider such a microphone arrangement. However, using this microphone arrangement simplifies equations. According to (26) and (37),
we can intersect circle and hyperbola to find source
location, x and y. For intersection of circle and hyperbola, firstly we rewrite (42) and (43), respectively, as
xs x1 2 ys y1 2 r 21
53
and
xs x2 2 ys y2 2 r 22 :
54
55
x R2 y02 r 22 ;
56
and
therefore,
r 21 xR2 r 22 x R2
57
which results in
r 22 r 21 4Rx:
58
Page 8 of 19
59
x r 22 r 21 =4R
and
q
y r 21 xR2
60
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
Page 9 of 19
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
62
s1 t stT 1 s0 tT 01
63
s2 t stT 2 s0 tT 02 :
64
Page 10 of 19
where
R1s1 s2
tT 1 stT 2 dt
0
0
R2s1 s2
tT 1 s tT 2 dt
0
0
R3s1 s2
s tT 1 stT 2 dt
0
0
0
R4s1 s2
s tT 1 s tT 2 dt:
67
68
69
70
71
Also according to this fact that the signal and noise are
generated by independent sources, they are considered uncorrelated. Therefore, the noisy spectrum can be written as:
0
0
Rs1 s2
stT 1 s tT 1 stT 2
s0 tT 02 dt
65
S n w S w N w:
72
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
d n 0; 1; 2; 3 ::::
73
2
4
These notches will appear at the following frequencies:
f
c 2n 1:c
4d
74
Page 11 of 19
jH f j
10 log10 H 1 f 10
2
10
H 1 f
10 log10
:
H f
2
75
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
Page 12 of 19
v
sound 31
:
R
76
delay estimation between the first and third microphones (Figure 5). Multiplying the inverse function of
notch-filter in third microphone's spectrum leads to increase in accuracy. Now using r, , and , we can calculate source location x, y and z in 3D case:
8
< x r: cos : sin
y r: sin ; sin :
77
:
z r: cos
The reasons for choosing the shape of sphere for the
reflector are as follows [69]. The simplest type of
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
R
:
2 cos
78
79
Page 13 of 19
between the slit and the viewing screen and d is the slit
width, based on the assumption that L> > d (Fraunhofer
scattering), the wave intensity distribution observed (on
a screen) at an angle with respect to the incident direction is given by:
if k
:d
I sin2 k
;
sin
I
k2
80
:d
sin m;
81
m
m 0; 2; ::::
d
82
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
circular slits S1 and S2. The narrow circular slits split the
incident wave into two coherent waves. After passing
through the slits, the two waves spread out due to diffraction interfere with one another. If this transmitted
wave is made to fall on a screen some distance away, an
interference pattern of bright and dark fringes are observed on the screen, with the bright ones corresponding
to regions of maximum wave intensity while the dark
ones corresponding to those of minimum wave intensity.
As discussed before, for all points on the screen where
the path difference is some integer multiple of the wavelength, the two waves from the slits S1 and S2 arrive in
phase and bright fringes are observed. Thus, the condition for producing bright fringes is as in (82). Similarly,
dark fringes are produced on the screen if the two waves
arriving on the screen from slits S1 and S2 are exactly
out of phase. This happens if the path difference between the two waves is an odd integer multiple of halfwavelengths:
m 12 :
sin
m 0; 1; 2; ::::
d
83
Of course, this condition is changed using a halfsphere instead of a plane (Figure 7), where according to
the same distance R to the center, the two waves from
the slits S1 and S2 arrive in phase at the center and
bright fringes are observed there. This signal intensity
magnification is not suitable for the case of estimating
intensity level difference between two microphones
where one of them is covered by this reflector. However,
covering the third microphone by half-sphere reflector is
suitable where there is no need to estimate the intensity
level difference between two microphones 1 and 3.
Page 14 of 19
v
sound : 21
2R
and cos1
v
sound : 31
b.
vsound 20:05
p
273:15 TemperatureCentigrade
Figure 7 Intensity profile of the diffraction pattern resulting from a plane wave passing through two narrow circular slits in a half-sphere.
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
Page 15 of 19
Figure 8 Setup of the proposed 3D sound source localization method using four microphones.
c.
p
21 :vsound
21 :vsound : m
p and r 2
p
r1
1 m
1 m
d.
q
H 1 f
jHf j
10log10
H f
3
f.
if jHf j0 y
q
r 21 xR2
q
else y r 21 xR2
8
< xs r:cos:sin
ys r:sin:sin
:
zs r:cos
g. Go to 3.
Proposed method
results for angle of
arrival (degrees)
Absolute value
of error (degrees)
0.18
0.18
15
15.13
0.13
30
29.85
0.15
45
45.18
0.18
60
60.14
0.14
75
74.91
0.09
90
89.83
0.17
105
104.82
0.18
120
120.14
0.14
135
135.11
0.11
150
150.14
0.14
165
165.09
0.09
180
179.93
0.07
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
Proposed method
results for angle of
arrival (degrees)
Absolute value
of error (degrees)
10
10.18
0.18
5.16
0.16
0.05
0.05
15
14.15
0.15
30
29.91
0.09
45
45.19
0.19
60
60.13
0.13
75
75.11
0.11
90
89.82
0.18
accurately find the sound velocity in different temperatures. Angle of arrival calculations using downloaded
wideband quasi-periodic helicopter sound resulted in
Tables 1 and 2. Acquisition hardware was initialized to 96kHz sampling frequency and 16-bit resolution. Also signal
acquisition window length was set to 100 ms. A custom 2cm-diameter reflector (Figure 8) was placed behind the
third microphone. Using the aforementioned sound source,
the sound spectra shown in Figure 9 were measured using
the third microphone. Figure 9a shows the spectrum when
Page 16 of 19
the sound source has been placed behind the reflector and
Figure 9b when placed in front of it. Notches can be spotted in the second spectrum according to the reflector
diameter and (74). Finally, using our proposed TDE-ILDHRTF approach, first we tried to find that a dominant
source is active in localization area. Then, using (72), we
tried to reduce background noise and localize sound
source in 3D space. Table 3 shows the improved results
after the application of noise reduction method. Although
PHAT features high performance in noisy and reflective
environments, due to the rather constant and high valued
signal-to-reverberation ratio (SRR) in outdoor cases, this
feature could not be evaluated. We tried to find the effect
of probable reverberation on our localization system results. In order to alleviate reverberation, our experiments
indicated that at least a distance of 1.2 m with surrounding buildings would be necessary. Although the loudspeaker's power, the hardware set (microphone, preamplifier,
and sound card) amplification factor and the distance of
the microphones with the surrounding buildings indicate
the SNR value for reverberate signals in a real environment, this distance can be calculated, but it can also be indicated experimentally, which is easier. Tables 1 and 2
indicate that our implemented 3D sound source localization method features less than 0.2 degree error for angle
of arrival. Table 3 indicates less than 10% error for 3D
Figure 9 Recorded sound spectrum using the third microphone. (a) When the source has been placed behind the reflector; (b) when the
source has been placed in front of the reflector. The logarithmic (dB) graphs indicate spectral level in comparison to 16-bit full scale.
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
Page 17 of 19
Table 3 Results for proposed 3D sound source localization method (Section 7) using noise reduction procedure
Real location (m)
10
1.9
9.2
4.1
0.9
0.8
0.9
10
2.9
9.2
4.3
0.9
0.8
0.7
10
2.2
10.7
5.6
0.8
0.7
0.6
10
3.2
10.7
5.8
0.8
0.7
0.8
10
5.5
9.5
5.8
0.5
0.5
0.8
10
6.5
10.5
4.5
0.5
0.5
0.5
10
7.7
10.8
4.4
0.7
0.8
0.6
10
8.6
10.8
4.4
0.6
0.8
0.6
10
9.8
9.1
4.2
0.8
0.9
0.8
10
10
9.5
9.4
5.9
0.5
0.6
0.9
10
15
10.6
15.7
5.5
0.6
0.7
0.5
10
20
10.7
19.2
5.6
0.7
0.8
0.6
10
25
9.2
23.9
5.7
0.8
1.1
0.7
10
30
10.5
31.7
4.3
0.5
1.7
0.7
10
10.8
9.7
5.8
0.8
0.7
0.8
10
9.3
8.7
5.8
0.7
0.7
0.8
10
10.6
6.5
5.9
0.6
0.5
0.9
10
9.5
6.6
4.1
0.5
0.6
0.9
10
10.5
4.5
4.2
0.5
0.5
0.8
10
10.6
3.4
5.7
0.6
0.6
0.7
10
9.3
3.6
5.5
0.7
0.6
0.5
10
10.8
2.8
5.8
0.8
0.8
0.8
10
10.1
0.2
4.3
0.1
0.8
0.7
10
10.6
0.4
4.1
0.6
0.6
0.9
10
10.5
2.7
5.6
0.5
0.7
0.6
10
10.4
2.8
5.5
0.4
0.2
0.5
10
9.7
3.7
5.5
0.3
0.3
0.5
10
9.4
5.2
5.9
0.6
0.2
0.9
the angle of arrival, used either least-square error estimation or Kalman filtering techniques to predict motion path
of sound sources, in order to overcome their local calculation errors. Examples are [2], [13], [15], and [19]. Meanwhile, comparison of our results with these and other
works (results on approximately similar tasks), such as
[72] and [73], shows the approximately same accuracy nature of our proposed source location measurement
approach.
10. Conclusion
In this paper, we reported on the simulation of TDEILD-based 2D half-plane sound source localization using
only two microphones. Reduction of the microphone
count was our goal. Therefore, we also proposed and
implemented TDE-ILD-HRTF-based 3D entire-space sound
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
9.
10.
11.
12.
13.
14.
15.
16.
Authors information
AP was born in Azerbaijan. He has a Ph.D. in Electrical Engineering (Signal
Processing) from the Electrical Engineering Department, Amirkabir University
of Technology, Tehran, Iran. He is now an invited lecturer at the Electrical
Engineering Department, Amirkabir University of Technology and has been
teaching several courses (C++ programming, multimedia systems, digital
signal processing, digital audio processing, and digital image processing). His
research interests include statistical signal processing and applications, digital
signal processing and applications, digital audio processing and applications
(acoustic modeling, speech coding, text to speech, audio watermarking and
steganography, sound source localization, determined and underdetermined
blind source separation and scene analyzing), digital image processing and
applications (image and video coding, image watermarking and
steganography, video watermarking and steganography, object tracking and
scene matching) as well as BCI.
SMA received his B.S. and M.S. degrees in Electronics from the Electrical
Engineering Department, Amirkabir University of Technology, Tehran, Iran, in
1984 and 1987, respectively, and his Ph.D. in Engineering from the University
of Cambridge, Cambridge, U.K., in 1996, in the field of speech processing.
Since 1988, he has been a Faculty Member at the Electrical Engineering
Department, Amirkabir University of Technology, where he is currently an
associate professor and teaches several courses and conducts research in
electronics and communications. His research interests include speech
processing, acoustic modeling, robust speech recognition, speaker
adaptation, speech enhancement as well as audio and speech watermarking.
Received: 24 June 2013 Accepted: 21 November 2013
Published: 6 December 2013
References
1. GC Carter, Time delay estimation for passive sonar signal processing. IEEE T
Acoust S. ASSP-29, 462470 (1981)
2. E Weinstein, Optimal source localization and tracking from passive array
measurements. IEEE T. Acoust. S. ASSP-30, 6976 (1982)
3. H Wang, P Chu, Voice source localization for automatic camera pointing
system in videoconferencing, in Proceedings of the ICASSP, New Paltz, NY,
1922, 1997
4. J Caffery Jr, G Stber, Subscriber location in CDMA cellular networks. IEEE
Trans. Veh. Technol. 47, 406416 (1998)
5. JB Tsui, Fundamentals of Global Positioning System Receivers (Wiley, New
York, 2000)
6. F Klein, Finding an earthquake's location with modern seismic networks
(Earthquake Hazards Program, Northern California, 2000)
7. Y Huang, J Benesty, GW Elko, RM Mersereau, Real-time passive source
localization: A practical linear-correction least-squares approach. IEEE T.
Audio P. 9, 943956 (2001)
8. VJM Michaud, F Rouat, J Letourneau, Robust Sound Source Localization Using
a Microphone Array on a Mobile Robot (in Proceedings of the IEEE/RSJ
International Conference on Intelligent Robots and Systems, vol. 2, Las
Vegas), pp. 12281233
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
35.
36.
37.
Page 18 of 19
Pourmohammad and Ahadi EURASIP Journal on Audio, Speech, and Music Processing 2013, 2013:27
http://asmp.eurasipjournals.com/content/2013/1/27
38. M Omologo, P Svaizer, Acoustic Event Localization using a CrosspowerSpectrum Phase based Technique (in Proceedings of the ICASSP, vol. 2,
Adelaide, 1994), pp. 273276
39. C Zhang, D Florencio, Z Zhang, Why Does PHAT Work Well in Low Noise,
reverberative Environments (in Proceedings of the ICASSP, Las Vegas, NV,
2008), pp. 25652568
40. W Cui, Z Cao, J Wei, DUAL-Microphone Source Location Method in 2-D Space
(in Proceedings of the ICASSP, Toulouse, 2006), pp. 845848
41. X Sheng, YH Hu, Maximum likelihood multiple-source localization using
acoustic energy measurements with wireless sensor networks. IEEE Trans
Signal Process 53(1), 4453 (2005)
42. D Blatt, AD Hero III, Energy-based sensor network source localization via
projection onto convex sets. IEEE Trans Signal Process 54(9), 36143619
(2006)
43. S Sen, A Nehorai, Performance analysis of 3-D direction estimation based on
head-related transfer function. IEEE Trans. Audio Speech Lang. Process
17(4) (2009)
44. DW Batteau, The role of the pinna in human localization. Series B, Biol. Sci
168, 158180 (1967)
45. SK Roffler, RA Butler, Factors that influence the localization of sound in the
vertical plane. J. Acoust. Soc. Am. 43(6), 12551259 (1968)
46. SR Oldfield, SPA Parker, Acuity of sound localization: a topography of
auditory space. Perception 13(5), 601617 (1984)
47. PM Hofman, JGAV Riswick, AJV Opstal, Relearing sound localization with
new ears. Nature Neurosci. 1(5), 417421 (1998)
48. CP Brown, Modeling the Elevation Characteristics of the Head-Related Impulse
Response (San Jose State University, San Jose, CA, M.S. Thesis, 1996)
49. P Satarzadeh, VR Algazi, RO Duda, Physical and filter pinna models based on
anthropometry (in Proceedings of the 122nd Conv, Audio Eng. Soc, Vienna,
2007), p. 7098
50. S Hwang, Y Park, Y Park, Sound Source Localization using HRTF database
(in ICCAS2005, Gyeonggi-Do, 2005)
51. WH Foy, Position-location solution by Taylor-series estimation. IEEE Trans.
Aerosp. Electron Syst. AES-12, 187194 (1976)
52. JS Abel, JO Smith, Source range and depth estimation from multipath
range difference measurements. IEEE Trans. Acoust. Speech. Signal. Process.
37, 11571165 (1989)
53. DMY Sommerville, Elements of Non-Euclidean Geometry (Dover, New York,
1958). pp. 2601958
54. LP Eisenhart, A Treatise on the Differential Geometry of Curves and Surfaces
(New York, Dover, 1960), pp. 270271. originally Ginn, 1909
55. JO Smith, JS Abel, Closed-Form least-squares source location estimation
from range-difference measurements. IEEE Trans. Acoust. Speech Signal
Process. ASSP-35, 16611669 (1987)
56. JM Delosme, M Morf, B Friedlander, A linear equation approach to locating
sources from time-difference-of-arrival measurements (in Proceedings of the
ICASSP, Denver, CO, 1980), pp. 0911
57. BT Fang, Simple solutions for hyperbolic and related position fixes. IEEE
Trans. Aerosp. Electron Syst. 26, 748753 (1990)
58. B Friedlander, A passive localization algorithm and its accuracy analysis. IEEE
J. Oceanic Eng. OE-12, 234245 (1987)
59. HC Schau, AZ Robinson, Passive source localization employing intersecting
spherical surfaces from time-of-arrival differences. IEEE Trans. Acoust. Speech
Signal Process. ASSP-35, 12231225 (1987)
60. JS Abel, A divide and conquer approach to least-squares estimation. IEEE
Trans. Aerosp. Electron. Syst. 26, 423427 (1990)
61. A Beck, P Stoica, J Li, Exact and approximate solutions of source localization
problems. IEEE Trans. Signal. Process. 56(5), 17701778 (2008)
62. HC So, YT Chan, FKW Chan, Closed-Form Formulae for Time-Difference-ofArrival Estimation. IEEE Trans Signal Process 56, 6 (2008)
63. MD Gillette, HF Silverman, A linear closed-form algorithm for source
localization from time-differences of arrival. IEEE Signal Process. Lett.
15, 14 (2008)
64. EG Larsson, D Danev, Accuracy comparison of LS and squared-range ls for
source localization. IEEE Trans. Signal. Process 58, 2 (2010)
65. N Ono, S Sagayama, R-means localization: a simple iterative algorithm for
range-difference-based source localization. in Proceedings of the ICASSP.
1419, 27182721 (March 2010)
66. ZE Chami, A Guerin, A Pham, C Servire, A PHASE-based Dual Microphone
Method to Count and Locate Audio Sources in Reverberant Rooms (in IEEE
67.
68.
69.
70.
71.
72.
73.
Page 19 of 19
doi:10.1186/1687-4722-2013-27
Cite this article as: Pourmohammad and Ahadi: N-dimensional Nmicrophone sound source localization. EURASIP Journal on Audio, Speech,
and Music Processing 2013 2013:27.