Documente Academic
Documente Profesional
Documente Cultură
multi-dimensional autocovariances
Fabian J. Theis1,2 , Anke Meyer-Baese2 , and Elmar W. Lang1
1
Institute of Biophysics,
University of Regensburg, D-93040 Regensburg, Germany
2
Department of Electrical and Computer Engineering,
Florida State University, Tallahassee, FL 32310-6046, USA
fabian@theis.name
Blind source separation (BSS) describes the task of recovering the unknown
mixing process and the underlying sources of an observed data set. Currently,
many BSS algorithm assume independence of the sources (ICA), see for instance
[1, 2] and references therein. In this work, we consider BSS algorithms based on
time-decorrelation. Such algorithms include AMUSE [3] and extensions such as
SOBI [4] and the similar TDSEP [5]. These algorithms rely on the fact that
the data sets have non-trivial autocorrelations. We give an extension thereof
to data sets, which have more than one direction in the parametrization, such
as images, by replacing one-dimensional autocovariances by multi-dimensional
autocovariances.
The paper is organized as follows: In section 1 we introduce the linear mixture
model; the next section 2 recalls results on time decorrelation BSS algorithms.
We then define multidimensional autocovariances and use them to propose mdSOBI in section 3. The paper finished with both artificial and real-world results
in section 4.
Linear BSS
We consider the following blind source separation (BSS) problem: Let x(t) be
an (observed) stationary m-dimensional real stochastical process (with not necessarily discrete time t) and A an invertible real matrix such that
x(t) = As(t) + n(t)
(1)
for all , and the additive noise n(t) is modelled by a stationary, temporally
and spatially white zero-mean process with variance 2 . x(t) is observed, and
the goal is to recover A and s(t). Having found A, s(t) can be estimated by
A1 x(t), which is optimal in the maximum-likelihood sense (if the density of
n(t) is maximal at 0, which is the case for usual noise models such as Gaussian or
Laplacian noise). So the BSS task reduces to the estimation of the mixing matrix
A. Extensions of the above model include for example the complex case [4] or the
allowance of different dimensions for s(t) and x(t), where the case of larger mixing
dimension can be easily reduced to the presented complete case by dimension
reduction resulting in a lower noise level [6].
By centering the processes, we can assume that x(t) and hence s(t) have zero
mean. The autocovariances then have the following structure
ARs (0)A> + 2 I
=0
>
(2)
Rx ( ) = E x(t + )x(t) =
ARs ( )A>
6= 0
Clearly, A (and hence s(t)) can be determined by equation 1 only up to permutation and scaling of columns. Since we assume existing variances of x(t)
and hence s(t), the scaling indeterminacy can be eliminated by the convention
Rs (0) = I. In order to guarantee identifiability of A except for permutation
from the above model, we have to additionally assume that there exists a delay
such that Rs ( ) has pairwise different eigenvalues (for a generalization see [4],
theorem 2). Then using the spectral theorem it is easy to see from equation 2
that A is determined uniquely by x(t) except for permutation.
Equation 2 also gives an indication of how to perform BSS i.e. how to recover A
(t) :=
from x(t). The usual first step consists of whitening the no-noise term x
As(t) of the observed mixtures x(t) using an invertible matrix V such that V
x(t)
has unit covariance. V can simply be estimated from x(t) by diagonalization of
the symmetric matrix Rx (0) = Rx (0) 2 I, provided that the noise variance
2 is known. If more signals than sources are observed, dimension reduction can
be performed in this step, and the noise level can be reduced [6].
In the following without loss of generality, we will therefore assume that
(t) = As(t) has unit covariance for each t. By assumption, s(t) also has
x
unit covariance, hence I = E As(t)s(t)> A> = ARs (0)A> = AA> so A is
(3)
Multidimensional SOBI
The goal of this work is to improve SOBI performance for random processes
with a higher dimensional parametrization i.e. for data sets where the random
processes s and x do not depend on a single variable t, but on multiple variables
(z1 , . . . , zM ). A typical example is a source data set, in which each component
si represents an image of size h w. Then M = 2 and samples of s are given at
z1 = 1, . . . , h, z2 = 1, . . . , w. Classically, s(z1 , z2 ) is transformed to s(t) by fixing
a mapping from the two-dimensional parameter set to the one-dimensional time
parametrization of s(t), for example by concatenating columns or rows in the
case of a finite number of samples. If the time structure of s(t) is not used, as
in all classical ICA algorithms in which i.i.d. samples are assumed, this choice
does not influence the result. However, in time-structure based algorithms such
as AMUSE and SOBI results can vary greatly depending on the choice of this
mapping, see figure 2.
Without loss of generality we again assume centered random vectors. Then
define the multidimensional covariance to be
Rs (1 , . . . , M ) := E s(z1 + 1 , . . . , zM + M )s(z1 , . . . , zM )>
0.8
0.6
0.4
0.2
PSfrag replacements
50
100
150
200
250
300
Fig. 1. Example of one- and two-dimensional autocovariance coefficient of the grayscale 128 128 Lena image after normalization to variance 1.
where the expectation is taken over (z1 , . . . , zM ). Rs (1 , . . . , M ) can be estimated given equidistant samples by replacing random variables by sample values
and expectations by sums as usual.
The advantage of using multidimensional autocovariances lies in the fact that
now the multidimensional structure of the data set can be used more explicitly.
For example, if row concatenation is used to construct s(t) from the images,
horizontal lines in the image will only give trivial contributions to the autocovariance (see examples in figure 2 and section 4). Figure 1 shows the oneand two-dimensional autocovariance of the Lena image for varying respectively (1 , 2 ) after normalization of the image to variance 1. Clearly, the twodimensional autocovariance does not decay as quickly with increasing radius as
the one-dimensional covariance. Only at multiples of the image height, the onedimensional autocovariance is significantly high i.e. captures image structure.
Our contribution consists of using multidimensional autocovariances for joint
diagonalization. We replace the BSS assumption of diagonal one-dimensional autocovariances by diagonal multi-dimensional autocovariances of the sources. Note
that also the multidimensional covariance satisfies the equation 2.
Again we as
(1)
(1)
4
SOBI
SOBI transposed images
mdSOBI
mdSOBI transposed images
I)
crosstalking error E1 (A,
3.5
2.5
1.5
0.5
PSfrag replacements
0
10
20
30
40
50
K
(b) performance comparison
60
70
Fig. 2. Comparison of SOBI and mdSOBI when applied to (unmixed) images from (a).
The plot (b) plots the number K of time lags versus the crosstalking error E1 of the
and the unit matrix I; here A
has been recovered by bot SOBI
recovered matrix A
and mdSOBI given the images from (a) respectively the transposed images.
after whitening of x(z1 , . . . , zK ). The joint diagonalizer then equals A except for
permutation, given the generalized identifiability conditions from [4], theorem
2. Therefore, also the identifiability result does not change, see [4]. In practice,
(k)
(k)
we choose the (1 , . . . , M ) with increasing modulus for increasing k, but with
(k)
the restriction 1 > 0 in order to avoid using the same autocovariances on the
diagonal of the matrix twice.
Often, data sets do not have any substantial long-distance autocorrelations,
but quite high multi-dimensional close-distance correlations (see figure 1). When
performing joint diagonalization, SOBI weighs each matrix equally strong, which
can deteriorate the performance for large K, see simulation in section 4.
Figure 2(a) shows an example, in which the images have considerable vertical structure, but rather random horizontal structure. Each of the two images
consists of a concatenation of stripes of two images. For visual purposes, we
chose the width of the stripes to be rather large with 16 pixels. According to
the previous discussion we expect one-dimensional algorithms such as AMUSE
and SOBI to perform well on the images, but badly (for number of time lags
16) on the transposed images. If we apply AMUSE with = 20 to the images, we get excellent performance with a low crosstalking error with the unit
matrix of 0.084; if we however apply AMUSE to the transposed images, the error is high with 1.1. This result is further confirmed by the comparison plot in
figure 2(b); mdSOBI performs equally well on the images and the transposed
A)
crosstalking error E1 (A,
PSfrag replacements
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
0.5
Fig. 3. SOBI and mdSOBI performance dependence on noise level . Plotted is the
with the real mixing matrix A. See
crosstalking error E1 of the recovered matrix A
text for more details.
Results
cc: 0.08
cc: 0.19
cc: 0.11
cc: 0.21
cc: 0.43
cc: 0.21
cc: 0.16
cc: 0.86
Fig. 4. mdSOBI fMRI analysis. The data was reduced to the first 8 principal components. (a) shows the recovered component maps (white points indicate values stronger
than 3 standard deviations), and (b) their time courses. mdSOBI was performed with
K = 32. Component 5 represents inner ventricles, component 6 the frontal eye fields.
Component 8 is the desired stimulus component, which is mainly active in the visual
cortex; its time-course closely follows the on-off stimulus (indicated by the gray boxes)
their crosscorrelation lies at cc = 0.86 with a delay of roughly 2 seconds induced
by the BOLD effect.
photic simulation periods with rest. Simulation and rest periods comprised 10
repetitions each, i.e. 30s. Resolution was 3 3 4 mm. The slices were oriented parallel to the calcarine fissure. Photic stimulation was performed using
an 8 Hz alternating checkerboard stimulus with a central fixation point and a
dark background with a central fixation point during the control periods. The
first scans were discarded for remaining saturation effects. Motion artifacts were
compensated by automatic image alignment (AIR, [11]).
BSS, mainly based on ICA, nowadays is a quite common tool in fMRI analysis
(see for example [12]). Here, we analyze the fMRI data set using spatial decorrelation as separation criterion. Figure 4 shows the performance of mdSOBI; see figure text for interpretation. Using only the first 8 principal components, mdSOBI
could recover the stimulus component as well as detect additional components.
When applying SOBI to the data set, it could not properly detect the stimulus
component but found two components with crosscorrelations cc = 0.81 and
0.84 with the stimulus time course.
Conclusion
We have proposed an extension called mdSOBI of SOBI for data sets with multidimensional parametrizations, such as images. Our main contribution lies in
replacing the one-dimensional autocovariances by multi-dimensional autocovariances. In both simulations and real-world applications mdSOBI outperforms
SOBI for these multidimensional structures.
In future work, we will show how to perform spatiotemporal BSS by jointly
diagonalizing both spatial and time autocovariance matrices. We plan on applying these results to fMRI analysis, where we also want to use three-dimensional
autocovariances for 3d-scans of the whole brain.
Acknowledgements
The authors would like to thank Dr. Dorothee Auer from the Max Planck Institute of Psychiatry in Munich, Germany, for providing the fMRI data, and Oliver
Lange from the Department of Clinical Radiology, Ludwig-Maximilian University, Munich, Germany, for data preprocessing and visualization. FT and EL
acknowledge partial financial support by the BMBF in the project ModKog.
References
1. Cichocki, A., Amari, S.: Adaptive blind signal and image processing. John Wiley
& Sons (2002)
2. Hyv
arinen, A., Karhunen, J., Oja, E.: Independent component analysis. John
Wiley & Sons (2001)
3. Tong, L., Liu, R.W., Soon, V., Huang, Y.F.: Indeterminacy and identifiability of
blind identification. IEEE Transactions on Circuits and Systems 38 (1991) 499509
4. Belouchrani, A., Meraim, K.A., Cardoso, J.F., Moulines, E.: A blind source separation technique based on second order statistics. IEEE Transactions on Signal
Processing 45 (1997) 434444
5. Ziehe, A., Mueller, K.R.: TDSEP an efficient algorithm for blind separation using
time structure. In Niklasson, L., Boden, M., Ziemke, T., eds.: Proc. of ICANN98,
Sk
ovde, Sweden, Springer Verlag, Berlin (1998) 675680
6. Joho, M., Mathis, H., Lamber, R.: Overdetermined blind source separation: using
more sensors than source signals in a noisy mixture. In: Proc. of ICA 2000, Helsinki,
Finland (2000) 8186
7. Cardoso, J.F., Souloumiac, A.: Jacobi angles for simultaneous diagonalization.
SIAM J. Mat. Anal. Appl. 17 (1995) 161164
8. Yeredor, A.: Non-orthogonal joint diagonalization in the leastsquares sense with
application in blind source separation. IEEE Trans. Signal Processing 50 (2002)
15451553
9. Ziehe, A., Laskov, P., Mueller, K.R., Nolte, G.: A linear least-squares algorithm
for joint diagonalization. In: Proc. of ICA 2003, Nara, Japan (2003) 469474
10. Cardoso, J.F., Souloumiac, A.: Blind beamforming for non gaussian signals. IEE
Proceedings - F 140 (1993) 362370
11. Woods, R., Cherry, S., Mazziotta, J.: Rapid automated algorithm for aligning
and reslicing pet images. Journal of Computer Assisted Tomography 16 (1992)
620633
12. McKeown, M., Jung, T., Makeig, S., Brown, G., Kindermann, S., Bell, A., Sejnowksi, T.: Analysis of fMRI data by blind separation into independent spatial
components. Human Brain Mapping 6 (1998) 160188