18.12.2012 Views

Myeloid Leukemia

Myeloid Leukemia

Myeloid Leukemia

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Classification of AML by DNA-Oligonucleotide Microarrays 229<br />

3.4.1.2. STATISTICAL SIGNIFICANCE<br />

Microarrays can measure the expression of thousands of genes to identify<br />

expression changes between different biological states. Methods based on conventional<br />

t-tests provide the probability (P) that a difference in gene expression<br />

occurred by chance. Although p < 0.01 is significant in the context of<br />

experiments designed to evaluate small numbers, a microarray experiment for<br />

more 10,000 genes would identify 100 genes by chance. Thus, methods are<br />

needed to determine the significance of these changes while accounting for the<br />

enormous number of genes. Commonly, to address the multiple testing problem,<br />

false-discovery rates (FDR) of genes are calculated according to a statistical<br />

method adapted specifically for microarrays (13). As it automatically takes<br />

into account the fact that thousands of genes are simultaneously being tested,<br />

the FDR is a widely accepted method to measure statistical significance in<br />

genome-wide studies. A measure of statistical significance called the q-value<br />

is associated with each tested feature. Similarly to the p-value, the q-value<br />

gives each measured gene its own individual measure of significance. Whereas<br />

the p-value is a measure of significance in terms of the false-positive rate, the<br />

q-value is a measure in terms of the false-discovery rate (FDR). In a microarray<br />

data set, the q-value of a particular feature is the expected proportion of false<br />

positives incurred when calling that feature significant. The q-values can be<br />

used as an exploratory guide for which features to investigate further, e.g.,<br />

through the use of pathway applications or classification engines.<br />

3.4.2. Estimation of Prediction Performance<br />

The generalization performance of the different algorithms can be estimated<br />

by performing cross-validation methods (CV). These methods are based on the<br />

idea that the most unbiased test of the predictive error is by applying it to data<br />

that were not used in the building of the initial predictive model. A common<br />

application is to partition a dataset into two parts—to fit the model on the first<br />

part, and to assess the predictive capability of that model on the second part.<br />

Depending on the CV method, the complete data set is split into different proportions<br />

of a training set and a test set. Each approach is performed to determine<br />

the accuracy, i.e., the probability of correct classification of a previously<br />

unknown sample.<br />

3.4.2.1. LEAVE-ONE-OUT CROSS-VALIDATION<br />

The leave-one-out cross-validation (LOOCV) method is one of several<br />

approaches to estimating how well a model that was trained on training data is<br />

going to perform on future as-yet-unseen data. LOOCV implies that one sample<br />

is excluded from the complete data set n and the remaining samples are used<br />

for training. This training and prediction process is repeated n times to include

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!