06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

In this paper we provide a framework of detection and localization of multiple similar shapes or object instances from an<br />

image based on shape matching. There are three challenges about the problem. The first is the basic shape matching<br />

problem about how to find the correspondence and transformation between two shapes; second how to match shapes under<br />

occlusion; and last how to recognize and locate all the matched shapes in the image. We solve these problems by using<br />

both graph partition and shape matching in a global optimization framework. A Hough-like collaborative voting is adopted,<br />

which provides a good initialization, data-driven information, and plays an important role in solving the partial matching<br />

problem due to occlusion. Experiments demonstrate the efficiency of our method.<br />

13:50-14:10, Paper ThBT1.2<br />

Bag of Hierarchical Co-Occurrence Features for Image Classification<br />

Kobayashi, Takumi, National Inst. of Advanced Industrial Science and<br />

Otsu, Nobuyuki, National Inst. of Advanced Industrial Science and<br />

We propose a bag-of-hierarchical-co-occurrence features method incorporating hierarchical structures for image classification.<br />

Local co-occurrences of visual words effectively characterize the spatial alignment of objects‘ components. The<br />

visual words are hierarchically constructed in the feature space, which helps us to extract higher-level words and to avoid<br />

quantization error in assigning the words to descriptors. For extracting descriptors, we employ two types of features hierarchically:<br />

narrow (local) descriptors, like SIFT [1], and broad descriptors based on co-occurrence features. The proposed<br />

method thus captures the co-occurrences of both small and large components. We conduct an experiment on image classification<br />

by applying the method to the Caltech 101 dataset and show the favorable performance of the proposed method.<br />

14:10-14:30, Paper ThBT1.3<br />

Person Detection using Temporal and Geometric Context with a Pan Tilt Zoom Camera<br />

Del Bimbo, Alberto, Univ. of Florence<br />

Lisanti, Giuseppe, Univ. of Florence<br />

Masi, Iacopo, Univ. of Florence<br />

Pernici, Federico, Univ. of Florence<br />

In this paper we present a system that integrates automatic camera geometry estimation and object detection from a Pan<br />

Tilt Zoom camera. We estimate camera pose with respect to a world scene plane in real-time and perform human detection<br />

exploiting the relative space-time context. Using camera self-localization, 2D object detections are clustered in a 3D world<br />

coordinate frame. Target scale inference is further exploited to reduce the number of false alarms and to increase also the<br />

detection rate in the final non-maximum suppression stage. Our integrated system applied on real-world data shows superior<br />

performance with respect to the standard detector used.<br />

14:30-14:50, Paper ThBT1.4<br />

Disparity Map Refinement for Video based Scene Change Detection using a Mobile Stereo Camera Platform<br />

Haberdar, Hakan, Univ. of Houston<br />

Shah, Shishir, Univ. of Houston<br />

This paper presents a novel disparity map refinement method and vision based surveillance framework for the task of detecting<br />

objects of interest in dynamic outdoor environments from two stereo video sequences taken at different times and<br />

from different viewing angles by a mobile camera platform. The proposed framework includes several steps, the first of<br />

which computes disparity maps of the same scene in two video sequences. Preliminary disparity images are refined based<br />

on estimated disparities in neighboring frames. Segmentation is performed to estimate ground planes, which in turn are<br />

used for establishing spatial registration between the two video sequences. Finally, the regions of change are detected<br />

using the combination of texture and intensity gradient features. We present experiments on detection of objects of different<br />

sizes and textures in real videos.<br />

14:50-15:10, Paper ThBT1.5<br />

Using Symmetry to Select Fixation Points for Segmentation<br />

Kootstra, Gert, KTH<br />

Bergström, Niklas, Royal Inst. of Tech.<br />

Kragic, Danica, KTH<br />

- 280 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!