20.02.2019 Views

CLC-Conference-Proceeding-2018

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

As is usual in Mathematics, when dealing<br />

with a non-linear problem, the possibility of<br />

linearizing is analyzed. A very natural idea<br />

emerges that gives rise to the Alternating Least<br />

Squares xxiii . The idea is to initialize one of the<br />

matrices, and this would be related to the third<br />

question of 1.1, and optimize with respect to the<br />

other.<br />

The problem is solved considering the<br />

other matrix as the variable and in the next step<br />

the calculated solution is taken as initialization<br />

for the matrix and optimized with respect to the<br />

first matrix. This is an alternating procedure and<br />

the advantage is that each step is solving a linear<br />

Least Squares problem with inequality<br />

constraints (non negativity) with convex<br />

objective function for which there are methods<br />

such as, for example, Lawson and Hanson xxiv .<br />

The general idea of the ALS is the following xxv :<br />

Initialize the matrix A, A (0) randomly or using<br />

any other strategy.<br />

For k = 1,2, ..., until satisfying a stopping<br />

criterion<br />

Solve<br />

(1)<br />

Solve min‖Y − AX ( k) ‖; A ( k+1)<br />

(2) A≥0<br />

As it is observed, the solution of a non-linear<br />

optimization problem has been replaced by the<br />

solution in each step of two linear problems with<br />

constraints.<br />

Another way to describe this algorithm appears<br />

in xxvi and is as follows:<br />

Initialize A randomly or with any other strategy.<br />

Estimate X of the matrix equation A T AX = A T Y<br />

and solve the problem<br />

A.<br />

with fixed matrix<br />

Assign all the negative coefficients of X the<br />

value 0 or a chosen small positive value ε. 4.<br />

Estimate A of the matrix equation XX T A T =<br />

XY T solving the problem<br />

matrix X.<br />

with fixed<br />

5. Assign to all the negative coefficients of A the<br />

value 0 or a chosen small positive value ε.<br />

This procedure can be rewritten as follows: X ←<br />

max{ε, (A T A) −1 AY} = [A † Y]<br />

+<br />

A ← max{ε, YX T (XX T ) −1 AY} = [YX † ]+<br />

where A † is the Moore-Penrose pseudoinverse<br />

xxvii of A, ε is a small constant (often 10 −6 ) to<br />

force the coefficients to be positive.<br />

In the Alternating Least Squares,<br />

however, the convergence to a global minimum<br />

is not guaranteed, not even to a stationary point,<br />

only to a point at which the objective function<br />

stops decreasing. This procedure can be<br />

improved as can be seen in Chapter 4 of xxviii It is<br />

interesting to notice that the NMF can be<br />

considered as an extension of the Non-negative<br />

Least Squares (NLS), which is a particular case<br />

of the Least Squares with inequality constraints<br />

(LSI) as described below:<br />

Given a matrix A ∈ R IxJ and a data set y ∈ R I ,<br />

find a non-negative vector x ∈ R J that minimizes<br />

the cost function , that is, ,<br />

, subject to x ≥ 0.<br />

In xxix , the problem of Non-negative Least<br />

Squares is formulated as a particular case of<br />

Least Squares with inequality constraints (LSI).<br />

If the matrix A has full rank then we are in the<br />

presence of a strictly convex problem and the

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!