14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

1.3. Dependent component analysis 13<br />

1.3 Dependent component analysis<br />

In this section, we will discuss the relaxation of the BSS model by tak<strong>in</strong>g <strong>in</strong>to account additional<br />

structures <strong>in</strong> the data and dependencies between components. Many researchers have taken<br />

<strong>in</strong>terest <strong>in</strong> this generalization, which is crucial for the application <strong>in</strong> real-world sett<strong>in</strong>gs where<br />

such situations are to be expected.<br />

Here, we will consider model <strong>in</strong>determ<strong>in</strong>acies as well as actual separation algorithms. For<br />

the latter, we will employ a technique that has been the basis of one of the first ICA algorithms<br />

(Cardoso and Souloumiac, 1993), namely jo<strong>in</strong>t diagonalization (JD). It has s<strong>in</strong>ce become an<br />

important tool <strong>in</strong> ICA-based BSS and <strong>in</strong> BSS rely<strong>in</strong>g on second-order time-decorrelation (Belouchrani<br />

et al., 1997). Its task is, given a set of commut<strong>in</strong>g symmetric n × n matrices Ci, to<br />

f<strong>in</strong>d an orthogonal matrix A such that A ⊤ CiA is diagonal for all i. This generalizes eigenvalue<br />

decomposition (i = 1) and the generalized eigenvalue problem (i = 2), <strong>in</strong> which perfect<br />

factorization is always possible.<br />

Other extensions of the standard BSS model such as <strong>in</strong>clud<strong>in</strong>g s<strong>in</strong>gular matrices (Georgiev<br />

and Theis, 2004) will be omitted from the discussion <strong>in</strong> the follow<strong>in</strong>g.<br />

1.3.1 Algebraic BSS and multidimensional generalizations<br />

Consider<strong>in</strong>g the BSS model from equation (1.1)—or a more general, noisy version x(t) = As(t)+<br />

n(t)—the data can only be separated if we put additional conditions on the sources such as:<br />

• they are stochastically <strong>in</strong>dependent: ps(s1, . . . , sn) = ps1 (s1) · · · psn(sn),<br />

• each source is sparse i.e. it conta<strong>in</strong>s a certa<strong>in</strong> number of zeros or has a low p-norm for<br />

small p and fixed 2-norm,<br />

• s(t) is stationary and for all τ, it has diagonal autocovariances E(s(t + τ) s(t) ⊤ ); here<br />

zero-mean s(t) are assumed.<br />

In the follow<strong>in</strong>g, we will review BSS algorithms based on eigenvalue decomposition, JD and<br />

generalizations. Thereby, one of the above conditions is denoted by the term source condition,<br />

because we do not want to specialize on a s<strong>in</strong>gle model. The additive noise n(t) is modeled by<br />

a stationary, temporally and spatially white zero-mean process with variance σ 2 . Moreover, we<br />

will not deal with the more complicated underdeterm<strong>in</strong>ed case, so we assume that at most as<br />

many sources as sensors are to be extracted, i.e. n ≤ m.<br />

The signals x(t) are observed, and the goal is to recover A and s(t). Hav<strong>in</strong>g found A, s(t)<br />

can be estimated by A † x(t), which is optimal <strong>in</strong> the maximum-likelihood sense. Here † denotes<br />

the pseudo-<strong>in</strong>verse of A, which equals the <strong>in</strong>verse <strong>in</strong> the case of m = n. So the BSS task reduces<br />

to the estimation of the mix<strong>in</strong>g matrix A, hence the additive noise n is often neglected (after<br />

whiten<strong>in</strong>g). Note that <strong>in</strong> the follow<strong>in</strong>g we will assume that all signals are real-valued. Extensions<br />

to the complex case are straightforward.<br />

Approximate jo<strong>in</strong>t diagonalization<br />

Many BSS algorithms employ jo<strong>in</strong>t diagonalization (JD) techniques on some source condition<br />

matrices to identify the mix<strong>in</strong>g matrix. Given a set of symmetric matrices C := {C1, . . . , CK},

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!