06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

09:00-11:10, Paper ThAT9.7<br />

A Vision-Based Taiwanese Sign Language Recognition System<br />

Huang, Chung-Lin, National Tsing-Hua Univ.<br />

Tsai, Bo-Lin, National Tsing-Hua Univ.<br />

This paper presents a vision-based continuous sign language recognition system to interpret the Taiwanese Sign Language<br />

(TSL). The continuous sign language, which consists of a sequence of hold and movement segments, can be decomposed<br />

into non-signs and signs. The signs can be either static signs or dynamic signs. The former can be found in the hold<br />

segment, whereas the latter can be identified in the combination of hold and movement segments. We use Support Vector<br />

Machine (SVM) to recognize the static sign and apply HMM model to identify the dynamic signs. Finally, we use the<br />

finite state machine to verify the correctness of the grammar of the recognized TSL sentence, and correct the miss-recognized<br />

signs.<br />

09:00-11:10, Paper ThAT9.8<br />

Fusing Audio-Visual Nonverbal Cues to Detect Dominant People in Group Conversations<br />

Aran, Oya, Idiap Res. Inst.<br />

Gatica-Perez, Daniel,<br />

This paper addresses the multimodal nature of social dominance and presents multimodal fusion techniques to combine<br />

audio and visual nonverbal cues for dominance estimation in small group conversations. We combine the two modalities<br />

both at the feature extraction level and at the classifier level via score and rank level fusion. The classification is done by<br />

a simple rule-based estimator. We perform experiments on a new 10-hour dataset derived from the popular AMI meeting<br />

corpus. We objectively evaluate the performance of each modality and each cue alone and in combination. Our results<br />

show that the combination of audio and visual cues is necessary to achieve the best performance.<br />

09:00-11:10, Paper ThAT9.9<br />

Wavelet Domain Local Binary Pattern Features for Writer Identification<br />

Du, Liang, Huazhong Univ. of Science and Tech.<br />

You, Xinge, Huazhong Univ. of Science and Tech.<br />

Xu, Huihui, Huazhong Univ. of Science and Tech.<br />

Gao, Zhifan, Huazhong Univ. of Science and Tech.<br />

Tang, Yuanyan, Hongkong Baptist University<br />

The representation of writing styles is a crucial step of writer identification schemes. However, the large intra-writer variance<br />

makes it a challenging task. Thus, a good feature of writing style plays a key role in writer identification. In this<br />

paper, we present a simple and effective feature for off-line, text-independent writer identification, namely wavelet domain<br />

local binary patterns (WD-LBP). Based on WD-LBP, a writer identification algorithm is developed. WD-LBP is able to<br />

capture the essence of characteristics of writer while ignoring the variations intrinsic to every single writer. Unlike other<br />

texture framework method, we do not assign any statistical distribution assumption to the proposed method. This prevent<br />

us from making any, possibly erroneous, assumptions about the handwritten image feature distributions. The experimental<br />

results show that the proposed writer identification method achieves high accuracy of identification and outperforms recent<br />

writer identification method such as wavelet-GGD model and Gabor filtering method.<br />

09:00-11:10, Paper ThAT9.10<br />

Audio-Visual Classification and Fusion of Spontaneous Affective Data in Likelihood Space<br />

Nicolaou, Mihalis, Imperial Coll.<br />

Gunes, Hatice, Imperial Coll.<br />

Pantic, Maja, Imperial Coll.<br />

This paper focuses on audio-visual (using facial expression, shoulder and audio cues) classification of spontaneous affect,<br />

utilising generative models for classification (i) in terms of Maximum Likelihood Classification with the assumption that<br />

the generative model structure in the classifier is correct, and (ii) Likelihood Space Classification with the assumption<br />

that the generative model structure in the classifier may be incorrect, and therefore, the classification performance can be<br />

improved by projecting the results of generative classifiers onto likelihood space, and then using discriminative classifiers.<br />

Experiments are conducted by utilising Hidden Markov Models for single cue classification, and 2 and 3-chain coupled<br />

Hidden Markov Models for fusing multiple cues and modalities. For discriminative classification, we utilise Support<br />

- 266 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!