Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
- TAGS
- abstract
- icpr
- icpr2010.org
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
and video frames were used to evaluate the performance of the proposed approach. Experimental results demonstrated the<br />
validity of our approach.<br />
13:50-14:10, Paper ThBT6.2<br />
Scene Text Extraction with Edge Constraint and Text Collinearity<br />
Lee, Seonghun, KAIST<br />
Cho, MinSu, KAIST<br />
Jung, Kyomin, KAIST<br />
Kim, Jin Hyung, KAIST<br />
In this paper, we propose a framework for isolating text regions from natural scene images. The main algorithm has two<br />
functions: it generates text region candidates, and it verifies of the label of the candidates (text or non-text). The text region<br />
candidates are generated through a modified K-means clustering algorithm, which references texture features, edge information<br />
and color information. The candidate labels are then verified in a global sense by the Markov Random Field model<br />
where collinearity weight is added as long as most texts are aligned. The proposed method achieves reasonable accuracy<br />
for text extraction from moderately difficult examples from the ICDAR 2003 database.<br />
14:10-14:30, Paper ThBT6.3<br />
Typographical Features for Scene Text Recognition<br />
Weinman, Jerod, Grinnell Coll.<br />
Scene text images feature an abundance of font style variety but a dearth of data in any given query. Recognition methods<br />
must be robust to this variety or adapt to the query data’s characteristics. To achieve this, we augment a semi-Markov<br />
model—-integrating character segmentation and recognition—-with a bigram model of character widths. Softly promoting<br />
segmentations that exhibit font metrics consistent with those learned from examples, we use the limited information available<br />
while avoiding error-prone direct estimates and hard constraints. Incorporating character width bigrams in this fashion<br />
improves recognition on low-resolution images of signs containing text in many fonts.<br />
14:30-14:50, Paper ThBT6.4<br />
A Visual Attention based Approach to Text Extraction<br />
Sun, Qiaoyu, HuaihaiInstitute of Tech.<br />
Lu, Yue, East China Normal Univ.<br />
Sun, Shiliang, East China Normal Univ.<br />
A visual attention based approach is proposed to extract texts from complicated background in camera-based images.<br />
First, it applies the simplified visual attention model to highlight the region of interest (ROI) in an input image and to<br />
yield a map, named the VA map, consisting of the ROIs. Second, an edge map of image containing the edge information<br />
of four directions is obtained by Sobel operators. Character areas are detected by connected component analysis and<br />
merged into candidate text regions. Finally, the VA map is employed to confirm the candidate text regions. The experimental<br />
results demonstrate that the proposed method can effectively extract text information and locate text regions contained in<br />
camera-based images. It is robust not only for font, size, color, language, space, alignment and complexity of background,<br />
but also for perspective distortion and skewed texts embedded in images.<br />
14:50-15:10, Paper ThBT6.5<br />
New Wavelet and Color Features for Text Detection in Video<br />
Palaiahnakote, Shivakumara, National Univ. of Singapore<br />
Phan, Trung Quy, National Univ. of Singapore<br />
Tan, Chew-Lim, National Univ. of Singapore<br />
Automatic text detection in video is an important task for efficient and accurate indexing and retrieval of multimedia data<br />
such as events identification, events boundary identification etc. This paper presents a new method comprising of wavelet<br />
decomposition and color features namely R, G and B. The wavelet decomposition is applied on three color bands separately<br />
to obtain three high frequency sub bands (LH, HL and HH) and then the average of the three sub bands for each color<br />
band is computed further to enhance the text pixels in video frame. To take advantage of wavelet and color information,<br />
we again take the average of the three average images (AoA) obtained by the former step to increase the gap between text<br />
- 287 -