12.07.2015 Views

Naive Credal Classifier 2: an extension of Naive Bayes for delivering ...

Naive Credal Classifier 2: an extension of Naive Bayes for delivering ...

Naive Credal Classifier 2: an extension of Naive Bayes for delivering ...

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationTest <strong>of</strong> domin<strong>an</strong>ce <strong>an</strong>d indeterminate classificationsDefinitionClass c 1 dominates c 2 if P(c 1 ) > P(c 2 ) in <strong>an</strong>y distribution <strong>of</strong> the posteriorcredal set.If no class dominates c 1 , then c 1 is non-dominated.If there are more non-dominated classes, NCC returns all <strong>of</strong> them: theclassification is indeterminate.NCC becomes indeterminate on the inst<strong>an</strong>ces whose classificationwould be prior-dependent with NBC.Indeterminate classifications pro<strong>of</strong>ed to be viable in real world casestudies (e.g., dementia diagnosis).<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationIncomplete data setsMost classifiers (including NBC) ignore missing data.This is correct only if data are missing-at-r<strong>an</strong>dom (MAR).It is not possible to test the MAR hypothesis on the incomplete data.However, ignoring Non-MAR missing data c<strong>an</strong> lead to unreliableconclusions.Missing data c<strong>an</strong> be MAR <strong>for</strong> some features but not <strong>for</strong> some others;or c<strong>an</strong> be MAR only in training <strong>an</strong>d not in testing (or vice versa).<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationNCC2: NCC with conservative treatment <strong>of</strong> missing data(I)NCC2 receives the declaration <strong>of</strong> which features have Non-MARmissing data <strong>an</strong>d at which stage (learning, testing or both).NCC2 ignores MAR missing data.NCC2 deals conservatively with Non-MAR missing data.Conservative treatment <strong>of</strong> missing data (learning set)All possible completions <strong>of</strong> missing data are seen as possible.A set <strong>of</strong> likelihoods is computed.A set <strong>of</strong> posteriors is computed from a set <strong>of</strong> priors <strong>an</strong>d a set <strong>of</strong>likelihoods.The conservative treatment <strong>of</strong> missing data c<strong>an</strong> generate additionalindeterminacy.<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationNCC2: NCC with conservative treatment <strong>of</strong> missing data(I)NCC2 receives the declaration <strong>of</strong> which features have Non-MARmissing data <strong>an</strong>d at which stage (learning, testing or both).NCC2 ignores MAR missing data.NCC2 deals conservatively with Non-MAR missing data.Conservative treatment <strong>of</strong> missing data (learning set)All possible completions <strong>of</strong> missing data are seen as possible.A set <strong>of</strong> likelihoods is computed.A set <strong>of</strong> posteriors is computed from a set <strong>of</strong> priors <strong>an</strong>d a set <strong>of</strong>likelihoods.The conservative treatment <strong>of</strong> missing data c<strong>an</strong> generate additionalindeterminacy.<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationNCC2: NCC with conservative treatment <strong>of</strong> missing data(II)Conservative treatment <strong>of</strong> missing data in the inst<strong>an</strong>ce to beclassifiedAll possible completions <strong>of</strong> missing data are seen as possible, thusgiving rise to several virtual inst<strong>an</strong>ces.Test <strong>of</strong> domin<strong>an</strong>ce: c 1 should dominate c 2 on all the virtual inst<strong>an</strong>ces.A procedure allows to find out the domin<strong>an</strong>ce relationships withoutactually building the virtual inst<strong>an</strong>ces.Conservative treatment <strong>of</strong> missing data in the inst<strong>an</strong>ce to classify c<strong>an</strong>generate additional indeterminacy.<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationNCC2: NCC with conservative treatment <strong>of</strong> missing data(II)Conservative treatment <strong>of</strong> missing data in the inst<strong>an</strong>ce to beclassifiedAll possible completions <strong>of</strong> missing data are seen as possible, thusgiving rise to several virtual inst<strong>an</strong>ces.Test <strong>of</strong> domin<strong>an</strong>ce: c 1 should dominate c 2 on all the virtual inst<strong>an</strong>ces.A procedure allows to find out the domin<strong>an</strong>ce relationships withoutactually building the virtual inst<strong>an</strong>ces.Conservative treatment <strong>of</strong> missing data in the inst<strong>an</strong>ce to classify c<strong>an</strong>generate additional indeterminacy.<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationWhat to expect from NCC2By adopting imprecise probabilities, NCC2 is designed to be robust to:prior specification, especially critical on small data sets;Non-MAR missing data, critical on incomplete data sets.However, excessive indeterminacy is undesirable.What to expect from indeterminate classificationsTo preserve NCC2 reliability, avoiding too strong conclusions (a singleclass) on doubtful inst<strong>an</strong>ces.To convey sensible in<strong>for</strong>mation, dropping unlikely classes.<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationWhat to expect from NCC2By adopting imprecise probabilities, NCC2 is designed to be robust to:prior specification, especially critical on small data sets;Non-MAR missing data, critical on incomplete data sets.However, excessive indeterminacy is undesirable.What to expect from indeterminate classificationsTo preserve NCC2 reliability, avoiding too strong conclusions (a singleclass) on doubtful inst<strong>an</strong>ces.To convey sensible in<strong>for</strong>mation, dropping unlikely classes.<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationIndicators <strong>of</strong> per<strong>for</strong>m<strong>an</strong>ce <strong>for</strong> NCC2NCC2determinacy (% <strong>of</strong> determinate classifications);single-accuracy (% <strong>of</strong> determ. classification that are accurate);set-accuracy (% <strong>of</strong> indeterm. classifications that contain the trueclass);size <strong>of</strong> indeterminate output, i.e., avg. number <strong>of</strong> classes returnedwhen indeterminate.<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationIndicators <strong>for</strong> comparing NBC <strong>an</strong>d NCC2NCC2 vs NBCNBC(NCC2 D): accuracy <strong>of</strong> NBC on inst<strong>an</strong>ces determinatelyclassified by NCC2.NBC(NCC2 I): accuracy <strong>of</strong> NBC on inst<strong>an</strong>ces indeterminatelyclassified by NCC2.We expect NBC(NCC2 D) > NBC(NCC I).<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationExperiments on 18 UCI data setsMAR setup: 5% missing data generated via a MAR mech<strong>an</strong>ism; allfeatures declared as MAR to NCC2.Non-MAR setup: 5% missing data generated via a Non-MARmech<strong>an</strong>ism; all features declared as Non-MAR to NCC2.Average NBC accuracy under both settings: 82%.NBC vs NCC2NBC (NCC2 D):85%(95%)NBC (NCC2 I ):36%(69%)On each data set <strong>an</strong>d setup:NBC(NCC2 D)>NBC(NCC2 I)NCC2determinacy: 95%(52%)single accuracy:85%(95%)set-accuracy: 85%(96%)imprecise output size:∼=33% <strong>of</strong> the classesIndeterminate classifications do preserve the reliability <strong>of</strong> NCC2!<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationExperiments on 18 UCI data setsMAR setup: 5% missing data generated via a MAR mech<strong>an</strong>ism; allfeatures declared as MAR to NCC2.Non-MAR setup: 5% missing data generated via a Non-MARmech<strong>an</strong>ism; all features declared as Non-MAR to NCC2.Average NBC accuracy under both settings: 82%.NBC vs NCC2NBC (NCC2 D):85%(95%)NBC (NCC2 I ):36%(69%)On each data set <strong>an</strong>d setup:NBC(NCC2 D)>NBC(NCC2 I)NCC2determinacy: 95%(52%)single accuracy:85%(95%)set-accuracy: 85%(96%)imprecise output size:∼=33% <strong>of</strong> the classesIndeterminate classifications do preserve the reliability <strong>of</strong> NCC2!<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationExperiments on 18 UCI data setsMAR setup: 5% missing data generated via a MAR mech<strong>an</strong>ism; allfeatures declared as MAR to NCC2.Non-MAR setup: 5% missing data generated via a Non-MARmech<strong>an</strong>ism; all features declared as Non-MAR to NCC2.Average NBC accuracy under both settings: 82%.NBC vs NCC2NBC (NCC2 D):85%(95%)NBC (NCC2 I ):36%(69%)On each data set <strong>an</strong>d setup:NBC(NCC2 D)>NBC(NCC2 I)NCC2determinacy: 95%(52%)single accuracy:85%(95%)set-accuracy: 85%(96%)imprecise output size:∼=33% <strong>of</strong> the classesIndeterminate classifications do preserve the reliability <strong>of</strong> NCC2!<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationExperiments on 18 UCI data setsMAR setup: 5% missing data generated via a MAR mech<strong>an</strong>ism; allfeatures declared as MAR to NCC2.Non-MAR setup: 5% missing data generated via a Non-MARmech<strong>an</strong>ism; all features declared as Non-MAR to NCC2.Average NBC accuracy under both settings: 82%.NBC vs NCC2NBC (NCC2 D):85%(95%)NBC (NCC2 I ):36%(69%)On each data set <strong>an</strong>d setup:NBC(NCC2 D)>NBC(NCC2 I)NCC2determinacy: 95%(52%)single accuracy:85%(95%)set-accuracy: 85%(96%)imprecise output size:∼=33% <strong>of</strong> the classesIndeterminate classifications do preserve the reliability <strong>of</strong> NCC2!<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationNBC Probabilities vs indeterminate classifications(MAR setup, average over all data sets)100%Avg. over 18 data setsAccuracy achieved by NBC80%60%40%20%DeterminacyNBC(NCC2 D)NBC(NCC2 I)0%60.00% 70.00% 80.00% 90.00% 100.00%Probability estimated by NBC <strong>for</strong> the class it returnsHigher posterior probability <strong>of</strong> NBC → higher NCC2 determinacy.At <strong>an</strong>y level <strong>of</strong> posterior probability, NBC(NCC2 D)>NBC(NCC2 I).Striking drop on the inst<strong>an</strong>ces classified confidently by NBC.<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon


Introducing NCC2 Experimental Results S<strong>of</strong>tware demonstrationSummaryNCC2 extends <strong>Naive</strong> <strong>Bayes</strong> to imprecise probabilities, to robustly dealwith small data sets <strong>an</strong>d missing data.NCC2 becomes indeterminate on inst<strong>an</strong>ces whose classification isdoubtful indeed.Indeterminate classifications preserve the classifier’ reliability whileconveying sensible in<strong>for</strong>mation.Bibliography, s<strong>of</strong>tware <strong>an</strong>d m<strong>an</strong>uals: seewww.idsia.ch/~giorgio/jncc2.htmlS<strong>of</strong>tware with GUI to arrive soon!<strong>Naive</strong> <strong>Credal</strong> <strong>Classifier</strong> 2G. Cor<strong>an</strong>i, M. Zaffalon

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!