Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
- TAGS
- abstract
- icpr
- icpr2010.org
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Answering to a query like when a particular document was printed is quite helpful in practice especially forensic purposes.<br />
This study attempts to develop a general framework that makes use of image processing and pattern recognition principles<br />
for ink age determination in printed documents. The approach, at first, computationally extracts a set of suitable color features<br />
and then analyzes them to properly associate them with ink age. Finally, a neural net is designed and trained to determine<br />
ages of unknown samples. The dataset used for the present experiment consists of the cover pages of LIFE<br />
magazines published in between 1930’s and 70’s (five decades). Test results show that a viable framework for involving<br />
machines in assisting human experts for determining age of printed documents.<br />
13:30-16:30, Paper WeBCT9.23<br />
Automatic Detection and Localization of Natural Scene Text in Video<br />
Huang, Xiaodong, Beijing Univ. of Posts and Telecommunications<br />
Ma, Huadong, Beijing Univ. of Posts and Telecommunications<br />
Video scene text contains semantic information and thus can contribute significantly to video indexing and summarization.<br />
However, most of the previous approaches to detecting scene text from videos experience difficulties in handling texts<br />
with various character size and text alignments. In this paper, we propose a novel algorithm of scene text detection and<br />
localization in video. Based on our observation that text character strokes show intensive edge details in the fixed orientation<br />
no matter what text alignment and size are, a stroke map is first generated. In the scene text detection, we extract<br />
the texture feature of stroke map to locate text lines. The detected scene text lines are accurately located by using Harris<br />
corners in the stroke map. Experimental results show that this approach is robust and can be effectively applied to scene<br />
text detection and localization in video.<br />
13:30-16:30, Paper WeBCT9.24<br />
High-Level Feature Extraction using SIFT GMMs and Audio Models<br />
Inoue, Nakamasa, Tokyo Inst. of Tech.<br />
Saito, Tatsuhiko, Tokyo Inst. of Tech.<br />
Shinoda, Koichi, Tokyo Inst. of Tech.<br />
Furui, Sadaoki,<br />
We propose a statistical framework for high-level feature extraction that uses SIFT Gaussian mixture models (GMMs)<br />
and audio models. SIFT features were extracted from all the image frames and modeled by a GMM. In addition, we used<br />
mel-frequency cepstral coefficients and ergodic hidden Markov models to detect high-level features in audio streams. The<br />
best result obtained by using SIFT GMMs in terms of mean average precision on the TRECVID 2009 corpus was 0.150<br />
and was improved to 0.164 by using audio information.<br />
13:30-16:30, Paper WeBCT9.25<br />
Pairwise Features for Human Action Recognition<br />
Ta, Anh Phuong, Univ. de Lyon, CNRS, INSA-Lyon, LIRIS<br />
Wolf, Christian, INSA de Lyon<br />
Lavoue, Guillaume, Univ. de Lyon, CNRS<br />
Baskurt, Atilla, LIRIS, INSA Lyon<br />
Jolion, Jolion, Univ. de Lyon<br />
Existing action recognition approaches mainly rely on the discriminative power of individual local descriptors extracted<br />
from spatio-temporal interest points (STIP), while the geometric relationships among the local features are ignored. This<br />
paper presents new features, called pairwise features (PWF), which encode both the appearance and the spatio-temporal<br />
relations of the local features for action recognition. First STIPs are extracted, then PWFs are constructed by grouping<br />
pairs of STIPs which are both close in space and close in time. We propose a combination of two code<strong>book</strong>s for video<br />
representation. Experiments on two standard human action datasets: the KTH dataset and the Weizmann dataset show that<br />
the proposed approach outperforms most existing methods.<br />
13:30-16:30, Paper WeBCT9.26<br />
Group Activity Recognition by Gaussian Processes Estimation<br />
Cheng, Zhongwei, Chinese Acad. of Sciences<br />
Qin, Lei, Chinese Acad. of Sciences<br />
- 233 -