14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

158 Chapter 10. IEEE TNN 16(4):992-996, 2005<br />

994 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 16, NO. 4, JULY 2005<br />

2) Normalizethecolumns ��Y � a IY FFFYxI of ˆI X �� a<br />

��a���� and set 4b0.<br />

Multiply each column �� by 0I if the first element of �� is negative.<br />

3) Cluster ��Y� aIY FFFYxI <strong>in</strong> � 0 I groups qIY FFFYq�CI such<br />

that for any � a IY FFFY�Y�� 0 �� ` 4YV�Y � P q�, and<br />

�� 0 �� !4 for any �Y � belong<strong>in</strong>g to different groups<br />

4) Choseany �� P q� and put —� a ��. Thematrix e with<br />

columns �—�� � �aI is an estimation of the mix<strong>in</strong>g matrix, up to<br />

permutation and scal<strong>in</strong>g.<br />

We should mention that the very sparse case <strong>in</strong> different sett<strong>in</strong>gs is<br />

already considered <strong>in</strong> the literature, but <strong>in</strong> more restrictive sense. In [6],<br />

theauthors supposethat thesupports of theFourier transform of any<br />

two source signals are disjo<strong>in</strong>t sets—a much more restrictive condition<br />

than our condition. In [1], theauthors supposethat for any sourcethere<br />

exists a time-frequency w<strong>in</strong>dow where only this source is nonzero and<br />

that the time-frequency transform of each source is not constant on any<br />

time-frequency w<strong>in</strong>dow. We would like to mention that their condition<br />

should <strong>in</strong>clude also the case when the the time-frequency transforms<br />

of any two sources are not proportional <strong>in</strong> any time-frequency w<strong>in</strong>dow.<br />

Such a quantitative condition (without frequency representation) is presented<br />

<strong>in</strong> our Theorem 2, condition ii).<br />

�aI<br />

Fig. 1. Orig<strong>in</strong>al images.<br />

Fig. 2. Mixed (observed) images.<br />

Fig. 3. Estimated normalized images us<strong>in</strong>g the estimated matrix. The<br />

signal-to-noiseratios with thesources from Fig. 1 are232, 239, and 228 dB,<br />

respectively.<br />

B. Identification of Sources<br />

Theorem 3: (Uniqueness of Sparse Representation): Let r bethe<br />

set of all � P � such that the l<strong>in</strong>ear system e� a � has a solution<br />

with at least � 0 � CIzero components. If e fulfills A1), then there<br />

exists a subset rH & rwith measure zero with respect to r, such<br />

that for every � Pr�rHthis system has no other solution with this<br />

property.<br />

�<br />

Proof: Obviously r is theunion of all<br />

� 0 I a@�3Aa@@�0<br />

These coefficients are uniquely determ<strong>in</strong>ed if ��� does not<br />

belong to the set rH with measure zero with respect to r<br />

2.3)<br />

(see Theorem 3);<br />

Constructthesolution �� a ƒ@XY�A:itconta<strong>in</strong>s!�Y� <strong>in</strong>theplace<br />

�� for � aIYFFFY�0 I, the other its components are zero.<br />

III. SCA<br />

IA3@�0�CIA3 hyperplanes, produced by tak<strong>in</strong>g the l<strong>in</strong>ear hull of every<br />

subsets of the columns of e with � 0 I elements. Let rH betheunion<br />

of all <strong>in</strong>tersections of any two such subspaces. Then rH has a measure<br />

zero <strong>in</strong> r and satisfies the conclusion of the theorem. Indeed, assume<br />

that � P r�rHand e� a e"� a �, where � and "� haveat least<br />

� 0 � CIzeros. S<strong>in</strong>ce � TP rHY� belongs to only one hyperplane<br />

produced as a l<strong>in</strong>ear hull of some � 0 I columns —� Y FFFY —� of<br />

e. It means that the vectors � and "� have � 0 � CIzeros <strong>in</strong> places<br />

with <strong>in</strong>dexes <strong>in</strong> �IY FFFY�����IY FFFY��0I�. Now from theequation<br />

e@�0"�A aHit follows that the �0I vector columns —� Y FFFY —�<br />

of e are l<strong>in</strong>early dependent, which is a contradiction with A1).<br />

From Theorem 3 it follows that the sources are identifiable generically,<br />

i.e., up to a set with a measure zero, if they have level of sparse-<br />

In this section, we develop a method for the complete solution of the<br />

SCA problem. Now the conditions are formulated only <strong>in</strong> terms of the<br />

data matrix ˆ.<br />

Theorem 4: (SCA Conditions): Assumethat � � x and the<br />

matrix ˆ P<br />

ness grater than or equal to � 0�CI, and themix<strong>in</strong>g matrix is known.<br />

In the follow<strong>in</strong>g, we present an algorithm, based on the observation <strong>in</strong><br />

Theorem 3.<br />

Source Recovery Algorithm:<br />

1) Identify thetheset of �-codimensional subspaces r produced<br />

by tak<strong>in</strong>g the l<strong>in</strong>ear hull of every subsets of the columns of e<br />

with � 0 I elements.<br />

2) Repeat for � a 1toxX 2.1) Identify the space r Prconta<strong>in</strong><strong>in</strong>g �� Xa ˆ@XY�A, or,<strong>in</strong><br />

practical situation with presence of noise, identify the one to<br />

which thedistancefrom �� is m<strong>in</strong>imal and project �� onto<br />

r to ���.<br />

2.2) if r is produced by the l<strong>in</strong>ear hull of column vectors<br />

—� Y FFFY —� , then f<strong>in</strong>d coefficients !�Y� such that<br />

�0I<br />

��� a !�Y�—� X<br />

�2x satisfies the follow<strong>in</strong>g conditions:<br />

i) �<br />

thecolumns of ˆ lie<strong>in</strong> theunion r of different hy-<br />

� 0 I<br />

perplanes, each column lies <strong>in</strong> only one such hyperplane, each<br />

hyperplane conta<strong>in</strong>s at least � columns of ˆ such that each<br />

� 0 I of them are l<strong>in</strong>early <strong>in</strong>dependent;<br />

ii) � 0 I<br />

for each � P �IYFFFY�� there exist � a different<br />

� 0 P<br />

hyperplanes �r�Y�� �<br />

�aI <strong>in</strong> r such that their <strong>in</strong>tersection v� a<br />

’ �<br />

�aIr�Y� is 1-D subspace;<br />

iii) any � different v� span thewhole � .<br />

Then the matrix ˆ is representable uniquely (up to permutation and<br />

scal<strong>in</strong>g of thecolumns of e and rows of ƒ) <strong>in</strong> theform ˆ a eƒ,<br />

where the matrices e P �2� and ƒ P �2x satisfy theconditions<br />

A1) and A2), A3), respectively.<br />

Proof: Let v� bespanned by —� and set e a �—�� � �aI. Condition<br />

iii) implies that any hyperplane from r conta<strong>in</strong>s at most � 0 I vectors<br />

from e. By i) and ii), it follows that these vectors are exactly � 0 I:<br />

only <strong>in</strong> this case the calculation of the number of all hyperplanes by<br />

� 0 I<br />

�<br />

ii) will givethenumber <strong>in</strong> i): � a@� 0 IA a<br />

� 0 P � 0 I .Let<br />

e be a matrix whose column vectors are all vectors from e (taken <strong>in</strong><br />

an arbitrary order). S<strong>in</strong>ce every column vector � of ˆ lies only <strong>in</strong> one<br />

hyperplane from r, the l<strong>in</strong>ear system e� a � has uniquesolution,<br />

which has at least � 0 � CIzeros (see the Proof of Theorem 3). Let<br />

���� � �aI be � column vectors from ˆ, which span onehyperplane<br />

from r, and � 0 I of them are l<strong>in</strong>early <strong>in</strong>dependent (such vectors<br />

exist by i)). Then we have: e�� a ��, for some uniquely determ<strong>in</strong>ed<br />

vectors ��Y� aIYFFFY�0 I, which are l<strong>in</strong>early <strong>in</strong>dependent and have

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!