06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

We present a method of performing kernel space domain description of a dataset with incomplete entries without the need<br />

for imputation, allowing kernel features of a class of data with missing features to be rigorously described. This addresses<br />

the problem that absent data completion is usually required before kernel classifiers, such as support vector domain description<br />

(SVDD), can be applied; equally, few existing techniques for incomplete data adequately address the issue of<br />

kernel spaces. Our method, which we call instance-specific domain description (ISDD), uses a parametrisation framework<br />

to compute minimal kernelised distances between data points with missing features through a series of optimisation runs,<br />

allowing evaluation of the kernel distance while avoiding subjective completions of missing data. We compare results of<br />

our method against those achieved by SVDD applied to an imputed dataset, using synthetic and experimental datasets<br />

where feature absence has a non-trivial structure. We show that our methods can achieve tighter sphere bounds when applied<br />

to linear and quadratic kernels.<br />

13:30-16:30, Paper WeBCT8.12<br />

Learning a Strategy with Neural Approximated Temporal-Difference Methods in English Draughts<br />

Fausser, Stefan, Univ. of Ulm<br />

Schwenker, Friedhelm, Univ. of Ulm<br />

Having a large game-tree complexity and being EXPTIME-complete, English Draughts, recently weakly solved during<br />

almost two decades, is still hard to learn for intelligent computer agents. In this paper we present a Temporal-Difference<br />

method that is nonlinear neural approximated by a 4-layer multi-layer perceptron. We have built multiple English Draughts<br />

playing agents, each starting with a randomly initialized strategy, which use this method during self-play to improve their<br />

strategies. We show that the agents are learning by comparing their winning-quote relative to their parameters. Our best<br />

agent wins versus the computer draughts programs Neuro Draughts, KCheckers and CheckerBoard with the easych engine<br />

and looses to Chinook, GuiCheckers and CheckerBoard with the strong cake engine. Overall our best agent has reached<br />

an amateur league level.<br />

13:30-16:30, Paper WeBCT8.13<br />

Learning the Kernel Combination for Object Categorization<br />

Zhang, Deyuan, Harbin Inst. of Tech.<br />

Wang, Xiaolong, Harbin Inst. of Tech.<br />

Liu, Bingquan, Harbin Inst. of Tech.<br />

Although Support Vector Machines(SVM) succeed in classifying several image databases using image descriptors proposed<br />

in the literature, no single descriptor can be optimal for general object categorization. This paper describes a novel framework<br />

to learn the optimal combination of kernels corresponding to multiple image descriptors before SVM training, leading<br />

to solve a quadratic programming problem efficiently. Our framework takes into account the variation of kernel matrix<br />

and imbalanced dataset, which are common in real world image categorization tasks. Experimental results on Graz-01<br />

and Caltech-101 image databases show the effectiveness and robustness of our algorithm.<br />

13:30-16:30, Paper WeBCT8.14<br />

SemiCCA: Efficient Semi-Supervised Learning of Canonical Correlations<br />

Kimura, Akisato, NTT Corp.<br />

Kameoka, Hirokazu, NTT Corp.<br />

Sugiyama, Masashi, Tokyo Inst. of Tech.<br />

Nakano, Takuho, University of Tokyo<br />

Maeda, Eisaku, Communication Science Lab.<br />

Sakano, Hitoshi, NTT<br />

Ishiguro, Katsuhiko, NTT<br />

Canonical correlation analysis (CCA) is a powerful tool for analyzing multi-dimensional paired data. However, CCA tends<br />

to perform poorly when the number of paired samples is limited, which is often the case in practice. To cope with this<br />

problem, we propose a semi-supervised variant of CCA named “Semi CCA” that allows us to incorporate additional unpaired<br />

samples for mitigating overfitting. The proposed method smoothly bridges the eigenvalue problems of CCA and<br />

principal component analysis (PCA), and thus its solution can be computed efficiently just by solving a single (generalized)<br />

eigenvalue problem as the original CCA. Preliminary experiments with artificially generated samples and PASCAL VOC<br />

data sets demonstrate the effectiveness of the proposed method.<br />

- 215 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!