12.07.2015 Views

Smart User Models: Modelling the Humans in Ambient ... - UdG

Smart User Models: Modelling the Humans in Ambient ... - UdG

Smart User Models: Modelling the Humans in Ambient ... - UdG

SHOW MORE
SHOW LESS
  • No tags were found...

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

4 Gustavo González et al.methods for specialization. This methodology can be used to both, learn user featuresfrom user <strong>in</strong>formation stored <strong>in</strong> recommender systems and deliver <strong>the</strong> userfeatures to o<strong>the</strong>r recommender systems. For details on <strong>the</strong> SUM management,see [2].Therefore, user’s UMD for each application are def<strong>in</strong>ed by shift<strong>in</strong>g <strong>in</strong>formationfrom and to UMD’s of different exist<strong>in</strong>g doma<strong>in</strong>s accord<strong>in</strong>g to <strong>the</strong> weightedgraphs G (SUM, UMD i ) def<strong>in</strong>ed by each application where user <strong>in</strong>terplays.3 Support Vector Mach<strong>in</strong>es <strong>in</strong> <strong>User</strong> <strong>Modell<strong>in</strong>g</strong>The Support Vector Mach<strong>in</strong>e (SVM) is a type of learn<strong>in</strong>g mach<strong>in</strong>e for summariz<strong>in</strong>g<strong>in</strong>formation and modell<strong>in</strong>g from examples based on <strong>the</strong> statistical learn<strong>in</strong>g<strong>the</strong>ory, which implements <strong>the</strong> structural risk m<strong>in</strong>imization <strong>in</strong>ductive pr<strong>in</strong>ciple <strong>in</strong>order to obta<strong>in</strong> a good generalization from data sets of limited size [3,4]. Therehas been a great deal of research <strong>in</strong>terest <strong>in</strong> <strong>the</strong>se methods over <strong>the</strong> last years,because:– They provide good generalization on <strong>the</strong> data.– They are well suited for sparse data.– They exhibit <strong>in</strong>dependence of <strong>the</strong> results from <strong>the</strong> <strong>in</strong>put space dimension.Although <strong>in</strong>itially conceived for l<strong>in</strong>early separable two classes classificationproblems, new algorithms have already been derived to solve classification problemswith non-separable data, regression, ord<strong>in</strong>al regression, and multi-classproblems. Let T = {(x i , y i ) ; x i ∈ X , y i ∈ {−1, +1}} be a tra<strong>in</strong><strong>in</strong>g data set for ab<strong>in</strong>ary classification task, where classes are labelled as +1, -1. Let <strong>the</strong> decisionfunction based on a hyperplane be f(x) = sign(w · x + b). Accord<strong>in</strong>g to <strong>the</strong>statistical learn<strong>in</strong>g <strong>the</strong>ory, a good generalization is achieved by maximiz<strong>in</strong>g <strong>the</strong>marg<strong>in</strong> between <strong>the</strong> separat<strong>in</strong>g hyperplane, w · x + b = 0, and <strong>the</strong> closest datapo<strong>in</strong>ts for each class <strong>in</strong> <strong>the</strong> <strong>in</strong>put space. This optimal hyperplane can be determ<strong>in</strong>edby solv<strong>in</strong>g a quadratic programm<strong>in</strong>g problem. The decision function canthus be written as,)f (x) = sign( SV∑i=1α i y i (x i · x) + bIn order to expand <strong>the</strong> method to non-l<strong>in</strong>ear decision functions, <strong>the</strong> orig<strong>in</strong>al<strong>in</strong>put space, X , projects to ano<strong>the</strong>r higher dimension dot product space F, calledfeature space, via a nonl<strong>in</strong>ear map φ : X → F, with dim(F) >> dim(X ). In thisnew space <strong>the</strong> optimal hyperplane is derived. Denot<strong>in</strong>g <strong>the</strong> <strong>in</strong>ner product <strong>in</strong> F,(kernel) φ(x i ) · φ(x j ) = K(x i , x j ), <strong>the</strong> decision function is formulated <strong>in</strong> termsof this kernel.)f (x) = sign( SV∑i=1α i y i K(x i , x) + bAs an important consequence of <strong>the</strong> SVM procedure, just a few of <strong>the</strong> tra<strong>in</strong><strong>in</strong>gpatterns are significant for classification purposes, those hav<strong>in</strong>g a weight α i

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!