06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Answering to a query like when a particular document was printed is quite helpful in practice especially forensic purposes.<br />

This study attempts to develop a general framework that makes use of image processing and pattern recognition principles<br />

for ink age determination in printed documents. The approach, at first, computationally extracts a set of suitable color features<br />

and then analyzes them to properly associate them with ink age. Finally, a neural net is designed and trained to determine<br />

ages of unknown samples. The dataset used for the present experiment consists of the cover pages of LIFE<br />

magazines published in between 1930’s and 70’s (five decades). Test results show that a viable framework for involving<br />

machines in assisting human experts for determining age of printed documents.<br />

13:30-16:30, Paper WeBCT9.23<br />

Automatic Detection and Localization of Natural Scene Text in Video<br />

Huang, Xiaodong, Beijing Univ. of Posts and Telecommunications<br />

Ma, Huadong, Beijing Univ. of Posts and Telecommunications<br />

Video scene text contains semantic information and thus can contribute significantly to video indexing and summarization.<br />

However, most of the previous approaches to detecting scene text from videos experience difficulties in handling texts<br />

with various character size and text alignments. In this paper, we propose a novel algorithm of scene text detection and<br />

localization in video. Based on our observation that text character strokes show intensive edge details in the fixed orientation<br />

no matter what text alignment and size are, a stroke map is first generated. In the scene text detection, we extract<br />

the texture feature of stroke map to locate text lines. The detected scene text lines are accurately located by using Harris<br />

corners in the stroke map. Experimental results show that this approach is robust and can be effectively applied to scene<br />

text detection and localization in video.<br />

13:30-16:30, Paper WeBCT9.24<br />

High-Level Feature Extraction using SIFT GMMs and Audio Models<br />

Inoue, Nakamasa, Tokyo Inst. of Tech.<br />

Saito, Tatsuhiko, Tokyo Inst. of Tech.<br />

Shinoda, Koichi, Tokyo Inst. of Tech.<br />

Furui, Sadaoki,<br />

We propose a statistical framework for high-level feature extraction that uses SIFT Gaussian mixture models (GMMs)<br />

and audio models. SIFT features were extracted from all the image frames and modeled by a GMM. In addition, we used<br />

mel-frequency cepstral coefficients and ergodic hidden Markov models to detect high-level features in audio streams. The<br />

best result obtained by using SIFT GMMs in terms of mean average precision on the TRECVID 2009 corpus was 0.150<br />

and was improved to 0.164 by using audio information.<br />

13:30-16:30, Paper WeBCT9.25<br />

Pairwise Features for Human Action Recognition<br />

Ta, Anh Phuong, Univ. de Lyon, CNRS, INSA-Lyon, LIRIS<br />

Wolf, Christian, INSA de Lyon<br />

Lavoue, Guillaume, Univ. de Lyon, CNRS<br />

Baskurt, Atilla, LIRIS, INSA Lyon<br />

Jolion, Jolion, Univ. de Lyon<br />

Existing action recognition approaches mainly rely on the discriminative power of individual local descriptors extracted<br />

from spatio-temporal interest points (STIP), while the geometric relationships among the local features are ignored. This<br />

paper presents new features, called pairwise features (PWF), which encode both the appearance and the spatio-temporal<br />

relations of the local features for action recognition. First STIPs are extracted, then PWFs are constructed by grouping<br />

pairs of STIPs which are both close in space and close in time. We propose a combination of two code<strong>book</strong>s for video<br />

representation. Experiments on two standard human action datasets: the KTH dataset and the Weizmann dataset show that<br />

the proposed approach outperforms most existing methods.<br />

13:30-16:30, Paper WeBCT9.26<br />

Group Activity Recognition by Gaussian Processes Estimation<br />

Cheng, Zhongwei, Chinese Acad. of Sciences<br />

Qin, Lei, Chinese Acad. of Sciences<br />

- 233 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!