You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
Classification of AML by DNA-Oligonucleotide Microarrays 229<br />
3.4.1.2. STATISTICAL SIGNIFICANCE<br />
Microarrays can measure the expression of thousands of genes to identify<br />
expression changes between different biological states. Methods based on conventional<br />
t-tests provide the probability (P) that a difference in gene expression<br />
occurred by chance. Although p < 0.01 is significant in the context of<br />
experiments designed to evaluate small numbers, a microarray experiment for<br />
more 10,000 genes would identify 100 genes by chance. Thus, methods are<br />
needed to determine the significance of these changes while accounting for the<br />
enormous number of genes. Commonly, to address the multiple testing problem,<br />
false-discovery rates (FDR) of genes are calculated according to a statistical<br />
method adapted specifically for microarrays (13). As it automatically takes<br />
into account the fact that thousands of genes are simultaneously being tested,<br />
the FDR is a widely accepted method to measure statistical significance in<br />
genome-wide studies. A measure of statistical significance called the q-value<br />
is associated with each tested feature. Similarly to the p-value, the q-value<br />
gives each measured gene its own individual measure of significance. Whereas<br />
the p-value is a measure of significance in terms of the false-positive rate, the<br />
q-value is a measure in terms of the false-discovery rate (FDR). In a microarray<br />
data set, the q-value of a particular feature is the expected proportion of false<br />
positives incurred when calling that feature significant. The q-values can be<br />
used as an exploratory guide for which features to investigate further, e.g.,<br />
through the use of pathway applications or classification engines.<br />
3.4.2. Estimation of Prediction Performance<br />
The generalization performance of the different algorithms can be estimated<br />
by performing cross-validation methods (CV). These methods are based on the<br />
idea that the most unbiased test of the predictive error is by applying it to data<br />
that were not used in the building of the initial predictive model. A common<br />
application is to partition a dataset into two parts—to fit the model on the first<br />
part, and to assess the predictive capability of that model on the second part.<br />
Depending on the CV method, the complete data set is split into different proportions<br />
of a training set and a test set. Each approach is performed to determine<br />
the accuracy, i.e., the probability of correct classification of a previously<br />
unknown sample.<br />
3.4.2.1. LEAVE-ONE-OUT CROSS-VALIDATION<br />
The leave-one-out cross-validation (LOOCV) method is one of several<br />
approaches to estimating how well a model that was trained on training data is<br />
going to perform on future as-yet-unseen data. LOOCV implies that one sample<br />
is excluded from the complete data set n and the remaining samples are used<br />
for training. This training and prediction process is repeated n times to include