13.07.2015 Views

Thesis - Instituto de Telecomunicações

Thesis - Instituto de Telecomunicações

Thesis - Instituto de Telecomunicações

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

106 CHAPTER 5. FEATURE SELECTION AND CLASSIFICATION5.4 Uncertainty Based Classification and Classifier FusionIn this section we address a novel classification technique proposed in this thesis, to beapplied on data with low separability.We can assess the error probability of a MAP classification usingpe(x) = 1 − max g i (x), (5.14)iwhere the discriminant function g i (x) =p(w i |x) for the selected class is used to inferthe classification error probability of sample x in class i.However instead of g i (x), an estimate is used, g i (x) = ˆp(w i |x), based on a parametric ornonparametric mo<strong>de</strong>l, and learned from a training set. Two problems can occur: (1) we havetoo few data available and therefore poor estimates of the mo<strong>de</strong>l; (2) the mo<strong>de</strong>l does not fitthe data well (incorrect mo<strong>de</strong>l). These two problems are not reflected in the discriminantfunction; in spite of this, the discriminant function is used both for the classification <strong>de</strong>cisionand for the rejection <strong>de</strong>cision.We propose a new classification scheme based on the classification uncertainty, by evaluatingĝ i (x) as a random variable for a fixed x ∗ . Our intention is to be able to use informationfrom the classification error probability, even when the data is not well mo<strong>de</strong>led.We <strong>de</strong>ci<strong>de</strong>d to use the term uncertainty that is related to the measurement error [232],<strong>de</strong>fined by the International Organization for Standardization (ISO) as the dispersion ofthe values that could reasonably be attributed to the measurand [107].5.4.1 Uncertainty Mo<strong>de</strong>lingWe start by consi<strong>de</strong>ring a normal distribution estimation problem.maximum a posteriori (MAP) classifier, <strong>de</strong>fined by the <strong>de</strong>cision rule:Consi<strong>de</strong>r again the<strong>de</strong>ci<strong>de</strong> w i if i = argmax i (p(w i |x)). (5.15)We want to study the behavior of the error estimate of the MAP classifier ˆp ewi (x) =1 − ˆp(w i |x). We start by studying p(x|w i ), since p(w i |x) is given by

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!