27.03.2014 Views

SEKE 2012 Proceedings - Knowledge Systems Institute

SEKE 2012 Proceedings - Knowledge Systems Institute

SEKE 2012 Proceedings - Knowledge Systems Institute

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Where each<br />

is the influence vector<br />

for the -th node. We take the -norm on each so that it<br />

will encourage all the elements in to go to zero together.<br />

With the estimated , we can directly identify those most<br />

influential nodes. Let us denote the set of hub nodes by :<br />

<br />

FISTA Algorithm for SLIM<br />

Input: ,<br />

Initialization:<br />

Iterate for<br />

Step 1:<br />

, until convergence<br />

The cardinality of (i.e., the number of hub nodes) is<br />

controlled by the regularization parameter . When ,<br />

then all will become zero and none of the nodes will be<br />

selected as hub node (i.e., ). On the othe r hand, if<br />

, . Put anothe r way, the sm aller is, the<br />

more nodes will be selected as hub nodes. In practice, we<br />

could tune the regularizatio n parameter to achieve the<br />

desirable number of hub nodes.<br />

Although the formulation for SLIM is a co nvex<br />

optimization problem, the n on-smoothness arising from<br />

and the additional non-negativity constraint make th e<br />

computation quite challenging. In the next section, w e<br />

present an efficient and scalable solver for SLIM which<br />

could be applied to solve web-scale problems.<br />

Step 2:<br />

Step 3:<br />

Output:<br />

Since (4) is separable in ter ms of<br />

independently, i.e.,<br />

<br />

, we solve it by e ach<br />

<br />

IV. OPTIMIZATION FOR SLIM<br />

In this sectio n, we present an efficient optimization<br />

algorithm for solving SLIM in (3). We first introduce some<br />

necessary notations. Let<br />

be the<br />

smooth part of the objective in (3). The gradient of over<br />

takes the following form:<br />

<br />

In addition, is Lipschitz con tinuous with the<br />

constants , which is the maximum<br />

eigenvalue of . In other word, we have for any ,<br />

With these notations in place, we use the fast accelerated<br />

gradient method framework. In particular, we adopt the fastiterative<br />

shrinkage-thresholding algorithm (FISTA) [9] as<br />

shown in the next Algorithm.<br />

It is known that this algorithm has the optim al<br />

convergence rate of .<br />

To apply this algorithm, the main difficulty is how to<br />

compute the first step. Firstly, we observe Step 1 can be<br />

written as:<br />

<br />

<br />

The closed form solution of the above minimization<br />

problem can be characterized by the following Theorem.<br />

Theorem 1. The optimal solution for the following<br />

optimization can be represented as following,<br />

<br />

<br />

Let<br />

be indices for the positive values<br />

in and let be the complement of . Then the optimal<br />

is,<br />

Proof. Firstly, we utilize the fact that the dual norm of<br />

-norm is -norm, so that we could present as<br />

<br />

Therefore, (6) can be reformulated as:<br />

<br />

<br />

where .<br />

<br />

Interchange the min and max, we have,<br />

<br />

3

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!