120 6 <strong>Web</strong> Usage <strong>Mining</strong>P(s i |z k )=∑ m(s i , p j )P(z k |s i , p j )p j∈ ∈P∑ m(s ′s ′ i ∈S,p i , p j)P(z k |s ′ i , p j)j∈P∑(6.18)P(z k )= 1 m(s i , p j )P(zRk |s i , p j ) (6.19)s i ∈S,p j ∈Pwhere R = ∑ m(s i, , p j )s i ∈S,p j ∈PBasically, substituting equations (6.17)-(6.19) into (6.14) <strong>and</strong> (6.15) will result in themonotonically increasing of total likelihood Li of the observation data. The iterative implementationof the E-step <strong>and</strong> M-step is repeating until Li is converging to a local optimal limit,which means the calculated results can represent the optimal probability estimates of the usageobservation data. From the previous formulation, it is easily found that the computational complexityof the PLSA model is O(mnk), where m, n <strong>and</strong> k denote the number of user sessions,<strong>Web</strong> pages <strong>and</strong> latent factors, respectively.By now, we have obtained the conditional probability distribution of P(z k ), P(s i |z k ) <strong>and</strong>P(p j |z k ) by performing the E <strong>and</strong> M step iteratively. The estimated probability distributionwhich is corresponding to the local maximum likelihood contains the useful information forinferring semantic usage factors, performing <strong>Web</strong> user sessions clustering which are describedin next sections.6.2.2 Constructing User Access Pattern <strong>and</strong> Identifying Latent Factor withPLSAAs discussed in the previous section, note that each latent factor z k does really represent aspecific aspect associated with the usage co-occurrence activities in nature. In other words, foreach factor, there might exist a task-oriented user access pattern corresponding to it. We, thus,can utilize the class-conditional probability estimates generated by the PLSA model to producethe aggregated user profiles for characterizing user navigational behaviors. Conceptually, eachaggregated user profile will be expressed as a collection of pages, which are accompanied bytheir corresponding weights indicating the contributions to such user group made by thosepages. Furthermore, analyzing the generated user profile can lead to reveal common user accessinterests, such as dominant or secondary “theme” by sorting the page weights.Partitioning User SessionsFirstly, we begin with the probabilistic variable , which represents the occurrence probabilityin the condition of a latent class factor z k exhibited by a given user session s i . On the otherh<strong>and</strong>, the probabilistic distribution over the factor space of a specific user session s i can reflectthe specific user access preference over the whole latent factor space, therefore, it may be utilizedto uncover the dominant factors by distinguishing the top probability values. Therefore,for each user session s i , we can further compute a set of probabilities over the latent factorspace via Bayesian formula as follows:P(z k |s i )= P(s i|z k )P(z k )∑ P(s i |z k )P(z k )z k ∈Z(6.20)
6.2 <strong>Web</strong> Usage <strong>Mining</strong> using Probabilistic Latent Semantic Analysis 121Actually, the set of probabilities P(z k |s i ) is tending to be “sparse”, that is, for a given s i , typicallyonly few entries are significant different from predefined threshold. Hence we can classifythe user into corresponding cluster based on these probabilities greater than a given threshold.Since each user session can be expressed as a pages vector in the original n-dimensionalspace, we can create a mixture representation of the collection of user sessions within samecluster that associated with the factor z k in terms of a collection of weighted pages. The algorithmfor partitioning user session is described as follows.Algorithm 6.4: Partitioning user sessionsInput: A set of calculated probability values of P(z k |s i ), a user session-page matrix SP, <strong>and</strong> apredefined threshold μ.Output: A set of session clusters SCL =(SCL 1 ,SCL 2 ,···SCL k )Step 1: Set SCL 1 = SCL 2 = ···= SCL k = ϕ,Step 2: For each s i ∈ S, select P(z k |s i ),ifP(z k |s i ) ≥ μ, then SCL k = SCL k ∪ s i ,Step 3: If there are still users sessions to be clustered, go back to step 2,Step 4: Output session clusters SCL = {SCL k }.Characterizing Latent Semantic FactorAs mentioned in previous section, the core of the PLSA model is the latent factor space.From this point of view, how to characterize the factor space or explain the semantic meaningof factors is a crucial issue in PLSA model. Similarly, we can also utilize another obtainedconditional probability distribution by the PLSA model to identify the semantic meaning ofthe latent factor by partitioning <strong>Web</strong> pages into corresponding categories associated with thelatent factors.For each hidden factor z k , we may consider that the pages, whose conditional probabilitiesP(p j |z k ) are greater than a predefined threshold, can be viewed to provide similar functionalcomponents corresponding to the latent factor. In this way, we can select all pages with probabilitiesexc<strong>ee</strong>ding a certain threshold to form an topic-specific page group. By analyzing theURLs of the pages <strong>and</strong> their weights derived from the conditional probabilities, which areassociated with the specific factor, we may characterize <strong>and</strong> explain the semantic meaning ofeach factor. In next section, two examples with respect to the discovered latent factors arepresented. The algorithm to generating the topic-oriented <strong>Web</strong> page group is briefly describedas follows:Algorithm 6.5: Characterizing latent semantic factorsInput: A set of conditional probabilities, P ( p j |z k), a predefined threshold μ.Output: A set of latent semantic factors represented by several dominant pages.Step 1: Set PCL 1 = PCL 2 = ···= PCL k = ϕ ,Step 2: For each z k , choose all <strong>Web</strong> pages such that P ( p j |z k)≥ μ <strong>and</strong> P(zk∣ ∣ p j ) ≥ μ, thenconstruct PCL k = p j ∪ PCL k ,Step 3: If there are still pages to be classified, go back to step 2,Step 4: Output PCL = {PCL k }.
- Page 2 and 3:
Web Mining and Social Networking
- Page 4:
Guandong Xu • Yanchun Zhang • L
- Page 8 and 9:
VIIIPrefacefollowing characteristic
- Page 11:
Acknowledgements: We would like to
- Page 14 and 15:
XIVContents3.1.2 Basic Algorithms f
- Page 16 and 17:
XVIContentsPart III Social Networki
- Page 19:
Part IFoundation
- Page 22 and 23:
4 1 Introduction(3). Learning usefu
- Page 24 and 25:
6 1 Introductioncalled computationa
- Page 26 and 27:
8 1 Introduction• The data on the
- Page 28 and 29:
10 1 Introductionin a broad range t
- Page 31 and 32:
2Theoretical BackgroundsAs discusse
- Page 33 and 34:
2.2 Textual, Linkage and Usage Expr
- Page 35 and 36:
2.4 Eigenvector, Principal Eigenvec
- Page 37 and 38:
2.5 Singular Value Decomposition (S
- Page 39 and 40:
2.6 Tensor Expression and Decomposi
- Page 41 and 42:
2.7 Information Retrieval Performan
- Page 43 and 44:
2.8 Basic Concepts in Social Networ
- Page 45:
2.8 Basic Concepts in Social Networ
- Page 48 and 49:
30 3 Algorithms and TechniquesTable
- Page 50 and 51:
32 3 Algorithms and TechniquesSpeci
- Page 52 and 53:
34 3 Algorithms and Techniquesa sub
- Page 54 and 55:
36 3 Algorithms and TechniquesMetho
- Page 56 and 57:
38 3 Algorithms and TechniquesCusto
- Page 58 and 59:
40 3 Algorithms and TechniquesTable
- Page 60 and 61:
42 3 Algorithms and Techniquesa bSI
- Page 62 and 63:
44 3 Algorithms and Techniques{a}10
- Page 64 and 65:
46 3 Algorithms and Techniques3.2 S
- Page 66 and 67:
48 3 Algorithms and TechniquesConce
- Page 68 and 69:
50 3 Algorithms and TechniquesNaive
- Page 70 and 71:
52 3 Algorithms and Techniquesuses
- Page 72 and 73:
54 3 Algorithms and Techniquesin th
- Page 74 and 75:
56 3 Algorithms and Techniques// Fu
- Page 76 and 77:
58 3 Algorithms and Techniquesendd
- Page 78 and 79:
60 3 Algorithms and Techniquesstart
- Page 80 and 81:
62 3 Algorithms and TechniquesHere
- Page 82 and 83:
64 3 Algorithms and Techniques3.8.2
- Page 84 and 85:
66 3 Algorithms and Techniquesfor e
- Page 86 and 87:
68 3 Algorithms and Techniquesthat
- Page 89 and 90: 4Web Content MiningIn recent years
- Page 91 and 92: score(q,d)=4.2 Web Search 73V(q) ·
- Page 93 and 94: 4.2 Web Search 75algorithm. The Web
- Page 95 and 96: 4.3 Feature Enrichment of Short Tex
- Page 97 and 98: 4.4 Latent Semantic Indexing 794.4
- Page 99 and 100: Notation4.5 Automatic Topic Extract
- Page 101 and 102: 4.5 Automatic Topic Extraction from
- Page 103 and 104: 4.6 Opinion Search and Opinion Spam
- Page 105: 4.6 Opinion Search and Opinion Spam
- Page 108 and 109: 90 5 Web Linkage Mining5.2 Co-citat
- Page 110 and 111: 92 5 Web Linkage Mining{ /1 out deg
- Page 112 and 113: 94 5 Web Linkage Mininga =(a(1),·
- Page 114 and 115: 96 5 Web Linkage Mining5.4.1 Bipart
- Page 116 and 117: 98 5 Web Linkage MiningNext, consid
- Page 118 and 119: 100 5 Web Linkage Mining(5) Creatin
- Page 120 and 121: 102 5 Web Linkage Miningpower-law d
- Page 122 and 123: 104 5 Web Linkage MiningFig. 5.10.
- Page 124 and 125: 106 5 Web Linkage Miningbetween use
- Page 126 and 127: 6Web Usage MiningIn previous chapte
- Page 129 and 130: 6.1 Modeling Web User Interests usi
- Page 131 and 132: 6.1 Modeling Web User Interests usi
- Page 133 and 134: 6.1 Modeling Web User Interests usi
- Page 135 and 136: 6.1 Modeling Web User Interests usi
- Page 137: 6.2 Web Usage Mining using Probabil
- Page 141 and 142: 6.2 Web Usage Mining using Probabil
- Page 143 and 144: 6.3 Finding User Access Pattern via
- Page 145 and 146: 6.3 Finding User Access Pattern via
- Page 147 and 148: 6.3 Finding User Access Pattern via
- Page 149 and 150: 6.4 Co-Clustering Analysis of weblo
- Page 151 and 152: 6.5 Web Usage Mining Applications 1
- Page 153 and 154: 6.5 Web Usage Mining Applications 1
- Page 155 and 156: 6.5 Web Usage Mining Applications 1
- Page 157 and 158: 6.5 Web Usage Mining Applications 1
- Page 159 and 160: 6.5 Web Usage Mining Applications 1
- Page 161: Part IIISocial Networking and Web R
- Page 164 and 165: 146 7 Extracting and Analyzing Web
- Page 166 and 167: 148 7 Extracting and Analyzing Web
- Page 168 and 169: 150 7 Extracting and Analyzing Web
- Page 170 and 171: 152 7 Extracting and Analyzing Web
- Page 172 and 173: 154 7 Extracting and Analyzing Web
- Page 174 and 175: 156 7 Extracting and Analyzing Web
- Page 176 and 177: 158 7 Extracting and Analyzing Web
- Page 178 and 179: 160 7 Extracting and Analyzing Web
- Page 180 and 181: 162 7 Extracting and Analyzing Web
- Page 182 and 183: 164 7 Extracting and Analyzing Web
- Page 184 and 185: 166 7 Extracting and Analyzing Web
- Page 186 and 187: 168 7 Extracting and Analyzing Web
- Page 188 and 189:
170 8 Web Mining and Recommendation
- Page 190 and 191:
172 8 Web Mining and Recommendation
- Page 192 and 193:
174 8 Web Mining and Recommendation
- Page 194 and 195:
176 8 Web Mining and Recommendation
- Page 196 and 197:
178 8 Web Mining and Recommendation
- Page 198 and 199:
180 8 Web Mining and Recommendation
- Page 200 and 201:
182 8 Web Mining and Recommendation
- Page 202 and 203:
184 8 Web Mining and Recommendation
- Page 204 and 205:
186 8 Web Mining and Recommendation
- Page 206 and 207:
188 8 Web Mining and Recommendation
- Page 208 and 209:
190 9 Conclusionsries commonly used
- Page 210 and 211:
192 9 Conclusionsas computer scienc
- Page 212 and 213:
194 9 Conclusionsresearches have de
- Page 214 and 215:
196 References14. J. Ayres, J. Gehr
- Page 216 and 217:
198 References49. D. Chakrabarti, R
- Page 218 and 219:
200 References82. C. Dwork, R. Kuma
- Page 220 and 221:
202 References119. J. Hou and Y. Zh
- Page 222 and 223:
204 References151. A. N. Langville
- Page 224 and 225:
206 References186. J. K. Mui and K.
- Page 226 and 227:
208 References223. C. Shahabi, A. M
- Page 228:
210 References260. G.-R. Xue, D. Sh