Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
Abstract book (pdf) - ICPR 2010
- TAGS
- abstract
- icpr
- icpr2010.org
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
In this article, a new fragile, blind, high payload capacity, ROI (Region of Interest) preserving Medical image watermarking<br />
(MIW) technique in the spatial domain for gray scale medical images is proposed. We present a watermarking scheme<br />
that combines lossless data compression and encryption technique in application to medical images. The effectiveness of<br />
the proposed scheme, proven through experiments on various medical images through various image quality measure matrices<br />
such as PSNR, MSE and MSSIM enables us to argue that, the method will help to maintain Electronic Patient Report(EPR)/DICOM<br />
data privacy and medical image integrity.<br />
TuBT6 Topkapı Hall B<br />
Face Recognition – I Regular Session<br />
Session chair: Ross, Arun (West Virginia Univ.)<br />
13:0-13:50, Paper TuBT6.1<br />
Efficient Facial Attribute Recognition with a Spatial Code<strong>book</strong><br />
Ijiri, Yoshihisa, OMRON Corp.<br />
Lao, Shihong, OMRON Corp.<br />
Han, Tony X., Univ. of Missouri<br />
Murase, Hiroshi, Nagoya Univ.<br />
There is a large number of possible facial attributes such as hairstyle, with/without glasses, with/without mustache, etc.<br />
Considering large number of facial attributes and their combinations, it is difficult to build attributes classifiers for all possible<br />
combinations needed in various applications, especially at the designing stage. To tackle this important and challenging<br />
problem, we propose a novel efficient facial attributes recognition algorithm using a learned spatial code<strong>book</strong>.<br />
The Maximum Entropy and Maximum Orthogonality (MEMO) criterion is followed to learn the spatial code<strong>book</strong>. With<br />
a spatial code<strong>book</strong> constructed at the designing stage, attribute classifiers can be trained on demand with a small number<br />
of exemplars with high accuracy on the testing data. Meanwhile, up to 600 times speedup is achieved in the on-demand<br />
training process, compared to current state-of-the-art method. The effectiveness of the proposed method is supported by<br />
convincing experimental results.<br />
13:50-14:10, Paper TuBT6.2<br />
Feature Space Hausdorff Distance for Face Recognition<br />
Chen, Shaokang, NICTA<br />
Lovell, Brian Carrington, The Univ. of Queensland<br />
We propose a novel face image similarity measure based on Hausdorff distance (HD). In contrast to conventional HDbased<br />
measures, which are generally applied in the image space (such as edge maps or gradient images), the proposed<br />
HD-based similarity measure is applied in the feature space. By extending the concept of HD using a variable radius and<br />
reference set, we can generate a neighbourhood set for HD measures in feature space and then apply this concept for classification.<br />
Experiments on the Labeled Faces in the Wild; and FRGC datasets show that the proposed measure improves<br />
the overall classification performance quite dramatically, especially under the highly desirable low false acceptance rate<br />
conditions.<br />
14:10-14:30, Paper TuBT6.3<br />
How to Measure Biometric Information?<br />
Sutcu, Yagiz, Pol. Inst. of New York Univ.<br />
Sencar, Husrev Taha, TOBB Univ. of Ec. and Tech.<br />
Memon, Nasir, Pol. Inst. of New York Univ.<br />
Being able to measure the actual information content of biometrics is very important but also a challenging problem. Main<br />
difficulty here is not only related to the selected feature representation of the biometric data, but also related to the matching<br />
algorithm employed in biometric systems. In this paper, we propose a new measure for measuring biometric information<br />
using relative entropy between intra-user and inter-user distance distributions. As an example, we evaluated the proposed<br />
measure on a face image dataset.<br />
- 117 -