06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

TuAT1 Marmara Hall<br />

Object Detection and Recognition – I Regular Session<br />

Session chair: Jiang, Xiaoyi (Univ. of Münster)<br />

09:00-09:20, Paper TuAT1.1<br />

Learning an Efficient and Robust Graph Matching Procedure for Specific Object Recognition<br />

Revaud, Jerome, Univ. de Lyon, CNRS<br />

Lavoue, Guillaume, Univ. de Lyon, CNRS<br />

Ariki, Yasuo, Kobe Univ.<br />

Baskurt, Atilla, LIRIS, INSA Lyon<br />

We present a fast and robust graph matching approach for 2D specific object recognition in images. From a small number<br />

of training images, a model graph of the object to learn is automatically built. It contains its local key points as well as<br />

their spatial proximity relationships. Training is based on a selection of the most efficient subgraphs using the mutual information.<br />

The detection uses dynamic programming with a lattice and thus is very fast. Experiments demonstrate that<br />

the proposed method outperforms the specific object detectors of the state-of-the-art in realistic noise conditions.<br />

09:20-09:40, Paper TuAT1.2<br />

A New Biologically Inspired Feature for Scene Image Classification<br />

Jiang, Aiwen, Chinese Acad. of Sciences<br />

Wang, Chunheng, Chinese Acad. of Sciences<br />

Xiao, Baihua, Chinese Acad. of Sciences<br />

Dai, Ruvei, Chinese Acad. of Sciences<br />

Scene classification is a hot topic in pattern recognition and computer vision area. In this paper, based on the past research<br />

on vision neuroscience, we proposed a new biologically inspired feature method for scene image classification. The new<br />

feature accounts for the visual processing from simple cell to complex cell in V1 area, and also the spatial layout for scene<br />

gist signature. It provides a different line and model revision to consider some nonlinearities inV1 area. We compare it<br />

with traditional HMAX model and recently proposed ScSPM model, and experiment on a popular 15 scenes dataset. We<br />

show that our proposed method has many important differences and merits. The experiment results also show that our<br />

method outperforms the state-of-the-art like ScSPM and KSPM model.<br />

09:40-10:00, Paper TuAT1.3<br />

On a Quest for Image Descriptors based on Unsupervised Segmentation Maps<br />

Koniusz, Piotr, Univ. of Surrey<br />

Mikolajczyk, Krystian, Univ. of Surrey<br />

This paper investigates segmentation-based image descriptors for object category recognition. In contrast to commonly<br />

used interest points the proposed descriptors are extracted from pairs of adjacent regions given by a segmentation method.<br />

In this way we exploit semi-local structural information from the image. We propose to use the segments as spatial bins<br />

for descriptors of various image statistics based on gradient, colour and region shape. Proposed descriptors are validated<br />

on standard recognition benchmarks. Results show they outperform state-of-the-art reference descriptors with 5.6x less<br />

data and achieve comparable results to them with 8.6x less data. The proposed descriptors are complementary to SIFT<br />

and achieve state-of-the-art results when combined together within a kernel based classifier.<br />

10:00-10:20, Paper TuAT1.4<br />

An RST-Tolerant Shape Descriptor for Object Detection<br />

Su, Chih-Wen, Acad. Sinica<br />

Liao, Mark, Acad. Sinica, Taiwan<br />

Liang, Yu-Ming, Acad. Sinica<br />

Tyan, Hsiao-Rong, Chung Yuan Christian Univ.<br />

In this paper, we propose a new object detection method that does not need a learning mechanism. Given a hand-drawn<br />

model as a query, we can detect and locate objects that are similar to the query model in cluttered images. To ensure the<br />

invariance with respect to rotation, scaling, and translation (RST), high curvature points (HCPs) on edges are detected<br />

first. Each pair of HCPs is then used to determine a circular region and all edge pixels covered by the circular region are<br />

- 71 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!