Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
- TAGS
- abstract
- icpr
- icpr2010.org
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
In this paper we provide a framework of detection and localization of multiple similar shapes or object instances from an<br />
image based on shape matching. There are three challenges about the problem. The first is the basic shape matching<br />
problem about how to find the correspondence and transformation between two shapes; second how to match shapes under<br />
occlusion; and last how to recognize and locate all the matched shapes in the image. We solve these problems by using<br />
both graph partition and shape matching in a global optimization framework. A Hough-like collaborative voting is adopted,<br />
which provides a good initialization, data-driven information, and plays an important role in solving the partial matching<br />
problem due to occlusion. Experiments demonstrate the efficiency of our method.<br />
13:50-14:10, Paper ThBT1.2<br />
Bag of Hierarchical Co-Occurrence Features for Image Classification<br />
Kobayashi, Takumi, National Inst. of Advanced Industrial Science and<br />
Otsu, Nobuyuki, National Inst. of Advanced Industrial Science and<br />
We propose a bag-of-hierarchical-co-occurrence features method incorporating hierarchical structures for image classification.<br />
Local co-occurrences of visual words effectively characterize the spatial alignment of objects‘ components. The<br />
visual words are hierarchically constructed in the feature space, which helps us to extract higher-level words and to avoid<br />
quantization error in assigning the words to descriptors. For extracting descriptors, we employ two types of features hierarchically:<br />
narrow (local) descriptors, like SIFT [1], and broad descriptors based on co-occurrence features. The proposed<br />
method thus captures the co-occurrences of both small and large components. We conduct an experiment on image classification<br />
by applying the method to the Caltech 101 dataset and show the favorable performance of the proposed method.<br />
14:10-14:30, Paper ThBT1.3<br />
Person Detection using Temporal and Geometric Context with a Pan Tilt Zoom Camera<br />
Del Bimbo, Alberto, Univ. of Florence<br />
Lisanti, Giuseppe, Univ. of Florence<br />
Masi, Iacopo, Univ. of Florence<br />
Pernici, Federico, Univ. of Florence<br />
In this paper we present a system that integrates automatic camera geometry estimation and object detection from a Pan<br />
Tilt Zoom camera. We estimate camera pose with respect to a world scene plane in real-time and perform human detection<br />
exploiting the relative space-time context. Using camera self-localization, 2D object detections are clustered in a 3D world<br />
coordinate frame. Target scale inference is further exploited to reduce the number of false alarms and to increase also the<br />
detection rate in the final non-maximum suppression stage. Our integrated system applied on real-world data shows superior<br />
performance with respect to the standard detector used.<br />
14:30-14:50, Paper ThBT1.4<br />
Disparity Map Refinement for Video based Scene Change Detection using a Mobile Stereo Camera Platform<br />
Haberdar, Hakan, Univ. of Houston<br />
Shah, Shishir, Univ. of Houston<br />
This paper presents a novel disparity map refinement method and vision based surveillance framework for the task of detecting<br />
objects of interest in dynamic outdoor environments from two stereo video sequences taken at different times and<br />
from different viewing angles by a mobile camera platform. The proposed framework includes several steps, the first of<br />
which computes disparity maps of the same scene in two video sequences. Preliminary disparity images are refined based<br />
on estimated disparities in neighboring frames. Segmentation is performed to estimate ground planes, which in turn are<br />
used for establishing spatial registration between the two video sequences. Finally, the regions of change are detected<br />
using the combination of texture and intensity gradient features. We present experiments on detection of objects of different<br />
sizes and textures in real videos.<br />
14:50-15:10, Paper ThBT1.5<br />
Using Symmetry to Select Fixation Points for Segmentation<br />
Kootstra, Gert, KTH<br />
Bergström, Niklas, Royal Inst. of Tech.<br />
Kragic, Danica, KTH<br />
- 280 -