06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

and video frames were used to evaluate the performance of the proposed approach. Experimental results demonstrated the<br />

validity of our approach.<br />

13:50-14:10, Paper ThBT6.2<br />

Scene Text Extraction with Edge Constraint and Text Collinearity<br />

Lee, Seonghun, KAIST<br />

Cho, MinSu, KAIST<br />

Jung, Kyomin, KAIST<br />

Kim, Jin Hyung, KAIST<br />

In this paper, we propose a framework for isolating text regions from natural scene images. The main algorithm has two<br />

functions: it generates text region candidates, and it verifies of the label of the candidates (text or non-text). The text region<br />

candidates are generated through a modified K-means clustering algorithm, which references texture features, edge information<br />

and color information. The candidate labels are then verified in a global sense by the Markov Random Field model<br />

where collinearity weight is added as long as most texts are aligned. The proposed method achieves reasonable accuracy<br />

for text extraction from moderately difficult examples from the ICDAR 2003 database.<br />

14:10-14:30, Paper ThBT6.3<br />

Typographical Features for Scene Text Recognition<br />

Weinman, Jerod, Grinnell Coll.<br />

Scene text images feature an abundance of font style variety but a dearth of data in any given query. Recognition methods<br />

must be robust to this variety or adapt to the query data’s characteristics. To achieve this, we augment a semi-Markov<br />

model—-integrating character segmentation and recognition—-with a bigram model of character widths. Softly promoting<br />

segmentations that exhibit font metrics consistent with those learned from examples, we use the limited information available<br />

while avoiding error-prone direct estimates and hard constraints. Incorporating character width bigrams in this fashion<br />

improves recognition on low-resolution images of signs containing text in many fonts.<br />

14:30-14:50, Paper ThBT6.4<br />

A Visual Attention based Approach to Text Extraction<br />

Sun, Qiaoyu, HuaihaiInstitute of Tech.<br />

Lu, Yue, East China Normal Univ.<br />

Sun, Shiliang, East China Normal Univ.<br />

A visual attention based approach is proposed to extract texts from complicated background in camera-based images.<br />

First, it applies the simplified visual attention model to highlight the region of interest (ROI) in an input image and to<br />

yield a map, named the VA map, consisting of the ROIs. Second, an edge map of image containing the edge information<br />

of four directions is obtained by Sobel operators. Character areas are detected by connected component analysis and<br />

merged into candidate text regions. Finally, the VA map is employed to confirm the candidate text regions. The experimental<br />

results demonstrate that the proposed method can effectively extract text information and locate text regions contained in<br />

camera-based images. It is robust not only for font, size, color, language, space, alignment and complexity of background,<br />

but also for perspective distortion and skewed texts embedded in images.<br />

14:50-15:10, Paper ThBT6.5<br />

New Wavelet and Color Features for Text Detection in Video<br />

Palaiahnakote, Shivakumara, National Univ. of Singapore<br />

Phan, Trung Quy, National Univ. of Singapore<br />

Tan, Chew-Lim, National Univ. of Singapore<br />

Automatic text detection in video is an important task for efficient and accurate indexing and retrieval of multimedia data<br />

such as events identification, events boundary identification etc. This paper presents a new method comprising of wavelet<br />

decomposition and color features namely R, G and B. The wavelet decomposition is applied on three color bands separately<br />

to obtain three high frequency sub bands (LH, HL and HH) and then the average of the three sub bands for each color<br />

band is computed further to enhance the text pixels in video frame. To take advantage of wavelet and color information,<br />

we again take the average of the three average images (AoA) obtained by the former step to increase the gap between text<br />

- 287 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!