06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

similarities among face carvings are computed by performing Procrustes analysis on local facial features (eyes, nose,<br />

mouth, etc.). The distance between corresponding face features is computed using point distribution models; the final pairwise<br />

similarity is the weighted sum of feature similarities. A web-based interface is provided to allow domain experts to<br />

interactively assign different weights to each face feature, and display hierarchical clustering results in 2D or 3D projections<br />

obtained by multidimensional scaling. The proposed framework has been successfully applied to the devata goddesses<br />

depicted in the ancient Angkor Wat temple. The resulting clusterings and visualization will enable a systematic anthropological,<br />

ethnological and artistic analysis of nearly 1,800 stone portraits of devatas of Angkor Wat.<br />

TuCT2 Topkapı Hall A<br />

Feature Extraction – II Regular Session<br />

Session chair: Covell, Michele (Google, Inc.)<br />

15:40-16:00, Paper TuCT2.1<br />

Action Recognition using Spatial-Temporal Context<br />

Hu, Qiong, Chinese Acad. of Sciences<br />

Qin, Lei, Chinese Acad. of Sciences<br />

Huang, Qingming, Chinese Acad. of Sciences<br />

Jiang, Shuqiang, Chinese Acad. of Sciences<br />

Tian, Qi, Univ. of Texas at San Antonio<br />

The spatial-temporal local features and the bag of words representation have been widely used in the action recognition<br />

field. However, this framework usually neglects the internal spatial-temporal relations between video-words, resulting in<br />

ambiguity in action recognition task, especially for videos in the wild. In this paper, we solve this problem by utilizing the<br />

volumetric context around a video-word. Here, a local histogram of video-words distribution is calculated, which is referred<br />

as the context and further clustered into contextual words. To effectively use the contextual information, the descriptive<br />

video-phrases (ST-DVPs) and the descriptive video-cliques (ST-DVCs) are proposed. A general framework for ST-DVP<br />

and ST-DVC generation is described, and then action recognition can be done based on all these representations and their<br />

combinations. The proposed method is evaluated on two challenging human action datasets: the KTH dataset and the<br />

YouTube dataset. Experiment results confirm the validity of our approach.<br />

16:00-16:20, Paper TuCT2.2<br />

Feature Extraction for Simple Classification<br />

Stuhlsatz, André, Univ. of Applied Sciences Duesseldorf<br />

Lippel, Jens, Univ. of Applied Sciences Duesseldorf<br />

Zielke, Thomas, Univ. of Applied Sciences Duesseldorf<br />

Constructing a recognition system based on raw measurements for different objects usually requires expert knowledge of<br />

domain specific data preprocessing, feature extraction, and classifier design. We seek to simplify this process in a way<br />

that can be applied without any knowledge about the data domain and the specific properties of different classification algorithms.<br />

That is, a recognition system should be simple to construct and simple to operate in practical applications. For<br />

this, we have developed a nonlinear feature extractor for high-dimensional complex patterns, using Deep Neural Networks<br />

(DNN). Trained partly supervised and unsupervised, the DNN effectively implements a nonlinear discriminant analysis<br />

based on a Fisher criterion in a feature space of very low dimensions. Our experiments show that the automatically extracted<br />

features work very well with simple linear discriminants, while the recognition rates improve only minimally if more sophisticated<br />

classification algorithms like Support Vector Machines (SVM) are used instead.<br />

16:20-16:40, Paper TuCT2.3<br />

Towards a Generic Feature-Selection Measure for Intrusion Detection<br />

Nguyen, Hai Thanh, Gjøvik Univ. Coll.<br />

Franke, Katrin, Gjøvik Univ. Coll.<br />

Petrovic, Slobodan, Gjøvik Univ. Coll.<br />

Performance of a pattern recognition system depends strongly on the employed feature-selection method. We perform an<br />

in-depth analysis of two main measures used in the filter model: the correlation-feature-selection (CFS) measure and the<br />

minimal-redundancy-maximal-relevance (mRMR) measure. We show that these measures can be fused and generalized<br />

into a generic feature-selection (GeFS) measure. Further on, we propose a new feature-selection method that ensures glob-<br />

- 121 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!