06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

In this paper, we present a calibration-free head gesture recognition system using a motion-sensor-based approach. For<br />

data acquisition we conducted a comprehensive study with 10 subjects. We analyzed the resulting head movement data<br />

with regard to separability and transferability to new subjects. Ordered means models (OMMs) were used for classification,<br />

since they provide an easy-to-use, fast, and stable approach to machine learning of time series. In result, we achieved classification<br />

rates of 85-95% for nodding, head shaking and tilting head gestures and good transferability. Finally, we show<br />

first promising attempts towards online recognition.<br />

09:00-11:10, Paper ThAT9.43<br />

TrajAlign: A Method for Precise Matching of 3-D Trajectories<br />

Aung, Zeyar, Inst. for Infocomm Res. Singapore<br />

Sim, Kelvin, Inst. for Infocomm Res. Singapore<br />

Ng, Wee Siong, Inst. for Infocomm Res. Singapore<br />

Matching two 3-D trajectories is an important task in a number of applications. The trajectory matching problem can be<br />

solved by aligning the two trajectories and taking the alignment score as their similarity measurement. In this paper, we<br />

propose a new method called “TrajAlign” (Trajectory Alignment). It aligns two trajectories by means of aligning their<br />

representative distance matrices. Experimental results show that our method is significantly more precise than the existing<br />

state-of-the-art methods. While the existing methods can provide correct answers in only up to 67% of the test cases, TrajAlign<br />

can offer correct results in 79% (i.e. 12% more) of the test cases, TrajAlign is also computationally inexpensive,<br />

and can be used practically for applications that demand efficiency.<br />

09:00-11:10, Paper ThAT9.44<br />

Real-Time 3D Model based Gesture Recognition for Multimedia Control<br />

Lin, Shih-Yao, National Taiwan Univ.<br />

Lai, Yun-Chien, National Taiwan Univ.<br />

Chan, Li-Wei, National Taiwan Univ.<br />

Hung, Yi-Ping, National Taiwan Univ.<br />

This paper presents a new 3D model-based gesture tracking system for controlling multimedia player in an intuitive way.<br />

The motivation of this paper is to make home appliance aware of user’s intention. This 3D model-based gesture tracking<br />

system adopts a Bayesian framework to track the user’s 3D hand position and to recognize meaning of these postures for<br />

controlling 3D player interactively. To avoid the high dimensionality of the whole 3D upper body model, which may complicate<br />

the gesture tracking problem, our system applies a novel hierarchical tracking algorithm to improve the system<br />

performance. Moreover, this system applies multiple cues for improving the accuracy of tracking results. Based on the<br />

above idea, we have implemented a 3D hand gesture interface for controlling multimedia players. Experimental results<br />

have shown that the proposed system robustly tracks the 3D position of the hand and has high potential for controlling the<br />

multimedia player.<br />

09:00-11:10, Paper ThAT9.45<br />

Motif Discovery and Feature Selection for CRF-Based Activity Recognition<br />

Zhao, Liyue, Univ. of Central Florida<br />

Wang, Xi, Univ. of Central Florida<br />

Sukthankar, Gita, Univ. of Central Florida<br />

Sukthankar, Rahul, Intel Labs Pittsburgh and Carnegie Mellon University<br />

Due to their ability to model sequential data without making unnecessary independence assumptions, conditional random<br />

fields (CRFs) have become an increasingly popular discriminative model for human activity recognition. However, how<br />

to represent signal sensor data to achieve the best classification performance within a CRF model is not obvious. This<br />

paper presents a framework for extracting motif features for CRF-based classification of IMU (inertial measurement unit)<br />

data. To do this, we convert the signal data into a set of motifs, approximately repeated symbolic sub sequences, for each<br />

dimension of IMU data. These motifs leverage structure in the data and serve as the basis to generate a large candidate set<br />

of features from the multi-dimensional raw data. By measuring reductions in the conditional log-likelihood error of the<br />

training samples, we can select features and train a CRF classifier to recognize human activities. An evaluation of our<br />

classifier on the CMU Multi-Modal Activity Database reveals that it outperforms the CRF-classifier trained on the raw<br />

features as well as other standard classifiers used in prior work.<br />

- 275 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!