Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
- TAGS
- abstract
- icpr
- icpr2010.org
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
13:30-16:30, Paper TuBCT8.32<br />
Action Detection in Crowded Videos using Masks<br />
Guo, Ping, Beijing Jiaotong Univ.<br />
Miao, Zhenjiang, Beijing Jiaotong Univ.<br />
In this paper, we investigate the task of human action detection in crowded videos. Different from action analysis in clean<br />
scenes, action detection in crowded environments is difficult due to the cluttered backgrounds, high densities of people<br />
and partial occlusions. This paper proposes a method for action detection based on masks. No human segmentation or<br />
tracking technique is required. To cope with the cluttered and crowded backgrounds, shape and motion templates are built<br />
and the shape templates are used as masks for feature refining. In order to handle the partial occlusion problem, only the<br />
moving body parts in each motion are involved in action training. Experiments using our approach are conducted on the<br />
CMU dataset with encouraging results.<br />
13:30-16:30, Paper TuBCT8.33<br />
3D Model based Vehicle Tracking using Gradient based Fitness Evaluation under Particle Filter Framework<br />
Zhang, Zhaoxiang, Beihang Univ.<br />
Huang, Kaiqi, Chinese Academy of Sciences<br />
Tan, Tieniu, Chinese Academy of Sciences<br />
Wang, Yunhong, Beihang Univ.<br />
We address the problem of 3D model based vehicle tracking from monocular videos of calibrated traffic scenes. A 3D<br />
wire-frame model is set up as prior information and an efficient fitness evaluation method based on image gradients is introduced<br />
to estimate the fitness score between the projection of vehicle model and image data, which is then combined<br />
into a particle filter based framework for robust vehicle tracking. Numerous experiments are conducted and experimental<br />
results demonstrate the effectiveness of our approach for accurate vehicle tracking and robustness to noise and occlusions.<br />
13:30-16:30, Paper TuBCT8.34<br />
Recovering 3D Shape and Light Source Positions from Non-Planar Shadows<br />
Yamashita, Yukihiro, Nagoya Inst. of Tech.<br />
Sakaue, Fumihiko, Nagoya Inst. of Tech.<br />
Sato, Jun, Nagoya Inst. of Tech.<br />
Recently, Shadow Graph has been proposed for recovering 3D shapes from shadows projected on curved surfaces. Unfortunately,<br />
this method requires a large computational cost. In this paper, we introduce 1D Shadow Graph which can be<br />
used for recovering 3D shapes with much smaller computational costs. We also extend our method, so that we can estimate<br />
both 3D shapes and light source positions simultaneously under a condition where 3D shapes and light sources are unknown.<br />
13:30-16:30, Paper TuBCT8.35<br />
3D Contour Model Creation for Stereo-Vision Systems<br />
Maruyama, Kenichi, National Inst. of Advanced Industrial Science and Tech.<br />
Kawai, Yoshihiro, National Inst. of Advanced Industrial Science and Tech.<br />
Tomita, Fumiaki, National Inst. of Advanced Industrial Science and Tech.<br />
The present paper describes a method for automatic 3D contour model creation for stereo-vision systems. The object<br />
model is a triangular surface mesh and a set of aspect models, which consists of model features and model points. Model<br />
features and model points are generated using 3D contours, which are estimated by the projected images of the triangular<br />
surface mesh from multiple discrete viewing directions. Using a non-photorealistic rendering approach, we extract not<br />
only the outer contours but also the inner contours of the projected images. Using both the inner and outer contours of the<br />
projected images, we create the object model which has 3D inner contour features and 3D contour generator features. Experimental<br />
results obtained using the 3D localization algorithm demonstrate the effectiveness of the proposed model.<br />
- 137 -