14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

1.3. Dependent component analysis 21<br />

ISA with known group structure via jo<strong>in</strong>t block diagonalization<br />

In order to solve ISA with fixed block-size k or at least known block structure, we will use a<br />

generalization of jo<strong>in</strong>t diagonalization, which searches for block-structures <strong>in</strong>stead of diagonality.<br />

We are not <strong>in</strong>terested <strong>in</strong> the order of the blocks, so the block-structure is uniquely specified by<br />

fix<strong>in</strong>g a partition n = m1 + . . . + mr of n and set m := (m1, . . . , mr) ∈ N r . An n × n matrix is<br />

said to be m-block diagonal if it is of the form<br />

⎛<br />

M1<br />

⎜<br />

⎝ .<br />

· · ·<br />

. ..<br />

0<br />

.<br />

⎞<br />

⎟<br />

⎠<br />

0 · · · Mr<br />

with arbitrary mi × mi matrices Mi.<br />

As generalization of JD <strong>in</strong> the case of known the block structure, the jo<strong>in</strong>t m-block diagonalization<br />

problem is def<strong>in</strong>ed as the m<strong>in</strong>imization of<br />

f m ( Â) :=<br />

K�<br />

�Â⊤Ci − diagm ( Â⊤CiÂ)�2F i=1<br />

(1.10)<br />

with respect to the orthogonal matrix Â, where diagm (M) produces a m-block diagonal matrix<br />

by sett<strong>in</strong>g all other elements of M to zero. Indeterm<strong>in</strong>acies of any m-JBD are m-scal<strong>in</strong>g<br />

i.e. multiplication by an m-block diagonal matrix from the right, and m-permutation def<strong>in</strong>ed<br />

by a permutation matrix that only swaps blocks of the same size.<br />

A few algorithms to actually perform JBD have been proposed, see Abed-Meraim and Belouchrani<br />

(2004), Févotte and Theis (2007a). In the follow<strong>in</strong>g we will simply perform jo<strong>in</strong>t<br />

diagonalization and then permute the columns of A to achieve block-diagonality—<strong>in</strong> experiments<br />

this turns out to be an efficient solution to JBD, although other more sophisticated pivot<br />

selection strategies for JBD are of <strong>in</strong>terest (Févotte and Theis, 2007b). The fact that JD <strong>in</strong>duces<br />

JBD has been conjectured by Abed-Meraim and Belouchrani (2004), and we were able to give<br />

a partial answer with the follow<strong>in</strong>g theorem:<br />

Theorem 1.3.2 (JBD via JD). Any block-optimal JBD of the Ci’s (i.e. a zero of f m ) is a local<br />

m<strong>in</strong>imum of the JD-cost-function f from equation (1.5).<br />

Clearly not any JBD m<strong>in</strong>imizes f, only those such that <strong>in</strong> each block of size mk, f( Â) when<br />

restricted to the block is maximal over A ∈ O(mk), which we denote as block-optimal. The<br />

proof is given <strong>in</strong> Theis (2007), see chapter 8.<br />

In the case of k-ISA, where m = (k, . . . , k), we used this result to propose an explicit<br />

algorithm (Theis, 2005a, see chapter 7). Consider the BSS model from equation (1.1). As usual<br />

by preprocess<strong>in</strong>g we may assume whitened observations x, so A is orthogonal. For the density<br />

ps of the sources, we therefore get ps(s0) = px(As0). Its Hessian transforms like a 2-tensor,<br />

which locally at s0 (see section 1.2.1) guarantees<br />

Hln ps (s0) = Hln px◦A(s0) = AHln px (As0)A ⊤ . (1.11)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!