06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

13:30-16:30, Paper WeBCT8.3<br />

Nonlinear Combination of Multiple Kernels for Support Vector Machines<br />

Li, Jinbo, East China Normal Univ.<br />

Sun, Shiliang, East China Normal Univ.<br />

Support vector machines (SVMs) are effective kernel methods to solve pattern recognition problems. Traditionally, they<br />

adopt a single kernel chosen beforehand, which makes them lack flexibility. The recent multiple kernel learning (MKL)<br />

overcomes this issue by optimizing over a linear combination of kernels. Despite its success, MKL neglects useful information<br />

generated from the nonlinear interaction of different kernels. In this paper, we propose SVMs based on the nonlinear<br />

combination of multiple kernels (NCMK) which surmounts the drawback of previous MKL by the potential to exploit<br />

more information. We show that our method can be formulated as a semi-definite programming (SDP) problem then solved<br />

by interior-point algorithms. Empirical studies on several data sets indicate that the presented approach is very effective.<br />

13:30-16:30, Paper WeBCT8.4<br />

Data Transformation of the Histogram Feature in Object Detection<br />

Zhang, Rongguo, Chinese Acad. of Sciences<br />

Xiao, Baihua, Chinese Acad. of Sciences<br />

Wang, Chunheng, Chinese Acad. of Sciences<br />

Detecting objects in images is very important for several application domains in computer vision. This paper presents an<br />

experimental study on data transformation of the feature vector in object detection. We use the modified Pyramid of Histograms<br />

of Orientation Gradients descriptor and the SVM classifier to form an object detection model. We apply a simple<br />

transformation to the histogram features before training and testing. This transformation equals a small change in the<br />

kernel function for Support Vector Machines. This change is much quicker than the kernel, but obtains better results. Experimental<br />

evaluations on the UIUC Image Database and TU Darmstadt Database show that the transformed features perform<br />

better than the raw features, and this transformation improves the linear separability of the histogram feature.<br />

13:30-16:30, Paper WeBCT8.5<br />

A New Learning Formulation for Kernel Classifier Design<br />

Sato, Atsushi, NEC<br />

This paper presents a new learning formulation for classifier design called ``General Loss Minimization.’’ The formulation<br />

is based on Bayes decision theory which can handle various losses as well as prior probabilities. A learning method for<br />

RBF kernel classifiers is derived based on the formulation. Experimental results reveal that the classification accuracy by<br />

the proposed method is almost the same as or better than Support Vector Machine (SVM), while the number of obtained<br />

reference vectors by the proposed method is much less than that of support vectors by SVM.<br />

13:30-16:30, Paper WeBCT8.6<br />

Variable Selection for Five-Minute Ahead Electricity Load Forecasting<br />

Koprinska, Irena, Univ. of Sydney<br />

Sood, Rohen, Univ. of Sydney<br />

Agelidis, Vassilios, Univ. of Sydney<br />

We use autocorrelation analysis to extract 6 nested feature sets of previous electricity loads for 5-minite ahead electricity<br />

load forecasting. We evaluate their predictive power using Australian electricity data. Our results show that the most important<br />

variables for accurate prediction are previous loads from the forecast day, 1, 2 and 7 days ago. By using also load<br />

variables from 3 and 6 days ago, we achieved small further improvements. The 3 bigger feature sets (37-51 features) when<br />

used with linear regression and support vector regression algorithms, were more accurate than the benchmarks. The overall<br />

best prediction model in terms of accuracy and training time was linear regression using the set of 51 features.<br />

13:30-16:30, Paper WeBCT8.7<br />

Enhancing Web Page Classification via Local Co-Training<br />

Du, Youtian, Xi’an Jiaotong Univ.<br />

Guan, Xiaohong, Xi’an Jiaotong Univ., Tsinghua University<br />

Cai, Zhongmin, Xi’an Jiaotong Univ.<br />

- 213 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!