06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

to obtain a background image. We have achieved human-area segmentation requiring no background image by using<br />

chamfer matching to match the results of human detection using Real AdaBoost with silhouette images. Although accuracy<br />

in chamfer matching drops as the number of templates increases, the proposed method enables segmentation accuracy to<br />

be improved by selecting silhouette images similar to the matching target beforehand based on response values from weak<br />

classifiers in Real AdaBoost.<br />

09:00-11:10, Paper ThAT8.7<br />

Local Optical Operators for Subpixel Scene Analysis<br />

Jean, Yves, City Univ. of NY<br />

In this paper we present a scene analysis technique with subpixel filtering based on dense coded light fields. Our technique<br />

computes alignment and optically projects analysis filters to local surfaces within the extent of a camera pixel. The resolution<br />

gain depends on the local light field density not on the point spread function of the camera optics. <strong>Abstract</strong> An initial<br />

structured light sequence is used in establishing each camera pixel’s footprint in the projector generated light field. Then<br />

a sequence of basis functions embedded in the light field, with camera pixel support, combine with the local surface texture<br />

and are integrated by a camera sensor to produce a localized response at the subpixel scale. We address optical modeling<br />

and aliasing issues since the dense light field is under sampled by the camera pixels. Results are provided with objects of<br />

planar and non-planar topology.<br />

09:00-11:10, Paper ThAT8.8<br />

Aesthetic Image Classification for Autonomous Agents<br />

Desnoyer, Mark, Carnegie Mellon Univ.<br />

Wettergreen, David, Carnegie Mellon Univ.<br />

Computational aesthetics is the study of applying machine learning techniques to identify aesthetically pleasing imagery.<br />

Prior work used online datasets scraped from large user communities like Flikr to get labeled data. However, online imagery<br />

represents results late in the media generation process, as the photographer has already framed the shot and then picked<br />

the best results to upload. Thus, this technique can only identify quality imagery once it has been taken. In contrast, automatically<br />

creating pleasing imagery requires understanding the imagery present earlier in the process. This paper applies<br />

computational aesthetics techniques to a novel dataset from earlier in that process in order to understand how the problem<br />

changes when an autonomous agent, like a robot or a real-time camera aid, creates pleasing imagery instead of simply<br />

identifying it.<br />

09:00-11:10, Paper ThAT8.9<br />

Removal of Moving Objects from a Street-view Image by Fusing Multiple Image Sequences<br />

Uchiyama, Hiroyuki, Nagoya Univ.<br />

Deguchi, Daisuke, Nagoya Univ.<br />

Takahashi, Tomokazu, Gifu Shotoku Gakuen Univ.<br />

Ide, Ichiro, Nagoya Univ.<br />

Murase, Hiroshi, Nagoya Univ.<br />

We propose a method to remove moving objects from an in-vehicle camera image sequence by fusing multiple image sequences.<br />

Driver assistance systems and services such as Google Street View require images containing no moving object.<br />

The proposed scheme consists of three parts: (i) collection of many image sequences along the same route by using vehicles<br />

equipped with an omni-directional camera, (ii) temporal and spatial registration of image sequences, and (iii) mosaicing<br />

partial images containing no moving object. Experimental results show that 97.3% of the moving object area could be removed<br />

by the proposed method.<br />

09:00-11:10, Paper ThAT8.10<br />

Improving SIFT-Based Descriptors Stability to Rotations<br />

Bellavia, Fabio, Univ. of Palermo<br />

Tegolo, Domenico, Univ. of Palermo<br />

Trucco, Emanuele<br />

Image descriptors are widely adopted structures to match image features. SIFT-based descriptors are collections of gradient<br />

- 251 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!