06.02.2013 Views

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

Abstract book (pdf) - ICPR 2010

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

10:00-10:20, Paper TuAT4.4<br />

Single Channel Speech Separation using Source-Filter Representation<br />

Stark, Michael, Graz Univ. of Tech.<br />

Wohlmayr, Michael, Graz Univ. of Tech.<br />

Pernkopf, Franz, Graz Univ. of Tech.<br />

We propose a fully probabilistic model for source-filter based single channel source separation. In particular, we perform<br />

separation in a sequential manner, where we estimate the source-driven aspects by a factorial HMM used for multi-pitch<br />

estimation. Afterwards, these pitch tracks are combined with the vocal tract filter model to form an utterance dependent<br />

model. Additionally, we introduce a gain estimation approach to enable adaptation to arbitrary mixing levels in the speech<br />

mixtures. We thoroughly evaluate this system and finally end up in a speaker independent model.<br />

10:20-10:40, Paper TuAT4.5<br />

Nonlinear Blind Source Separation using Slow Feature Analysis with Random Features<br />

Ma, Kuijun, Chinese Acad. of Sciences<br />

Tao, Qing, Chinese Acad. of Sciences<br />

Wang, Jue, Chinese Acad. of Sciences<br />

We develop an algorithm RSFA to perform nonlinear blind source separation with temporal constraints. The algorithm is<br />

based on slow feature analysis using random Fourier features for shift invariant kernels, followed by a selection procedure<br />

to obtain the sought-after signals. This method not only obtains remarkable results in a short computing time, but also excellently<br />

handles situations where there are multiple types of mixtures. In kernel methods, since the problem is unsupervised,<br />

the need of multiple kernels is ubiquitous. Experiments on music excerpts illustrate the strong performance of our<br />

method.<br />

TuAT5 Anadolu Auditorium<br />

Image Analysis – III Regular Session<br />

Session chair: Kittler, Josef (Univ. of Surrey)<br />

09:00-09:20, Paper TuAT5.1<br />

Canonical Image Selection by Visual Context Learning<br />

Zhou, Wengang, Univ. of Science and Tech. of China<br />

Lu, Yijuan, Texas State Univ. at San Marcos<br />

Li, Houqiang, Univ. of Science and Tech. of China<br />

Tian, Qi, Univ. of Texas at San Antonio<br />

Canonical image selection is to select a subset of photos that best summarize a photo collection. In this paper, we define<br />

the canonical image as those that contain most important and distinctive visual words. We propose to use visual context<br />

learning to discover visual word significance and develop Weighted Set Coverage algorithm to select canonical images<br />

containing distinctive visual words. Experiments with web image datasets demonstrate that the canonical images selected<br />

by our approach are not only representatives of the collected photos, but also exhibit a diverse set of views with minimal<br />

redundancy.<br />

09:20-09:40, Paper TuAT5.2<br />

Exposing Digital Image Forgeries by using Canonical Correlation Analysis<br />

Zhang, Chi, Beijing Univ. of Tech.<br />

Zhang, Hongbin, Beijing Univ. of Tech.<br />

In this paper, we propose a new method to detect the forgeries in digital images by using photo-response non-uniformity<br />

(PRNU) noise features. The method utilizes canonical correlation analysis (CCA) to measure linear correlation relationship<br />

between two sets of PRNU noise estimation from images taken by the same camera. The linear correlation relationship<br />

maximizes the correlation between the noise reference pattern(or PRNU noise estimation) and PRNU noise features from<br />

the same camera. To further improve the detection accuracy rate, the difference of variance between an image region and<br />

its smoothed version is used to categorize the image region into heavily textured region class or non-heavily textured<br />

region class. For a heavily textured region or a non-heavily textured region, Neyman-Pearson decision is used to calculate<br />

the corresponding threshold, and get the final result of detection.<br />

- 76 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!