11.07.2015 Views

A comparison of bootstrap methods and an adjusted bootstrap ...

A comparison of bootstrap methods and an adjusted bootstrap ...

A comparison of bootstrap methods and an adjusted bootstrap ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

form e = am −α+b with j=1, …, J. The estimates a , α <strong><strong>an</strong>d</strong> for the parameters aremjjobtained by minimizing the non-linear least squares functionJ∑j=1b { } 2mj−αje −am −b.The <strong>adjusted</strong> <strong>bootstrap</strong> estimate for the prediction error is given byABSe n = <strong>an</strong> −α+ b .It is the fitted value on the learning curve as if all original observations contributed to <strong>an</strong>individual <strong>bootstrap</strong> learning set.In practice, the choice <strong>of</strong> l c<strong>an</strong> r<strong>an</strong>ge from somewhere close to 1 to a value greater th<strong>an</strong> 5.Repeated leave-one-out <strong>bootstrap</strong> estimates typically have lower variability th<strong>an</strong> leaveone-outcross-validation. They have <strong>an</strong> upward bias that decreases <strong><strong>an</strong>d</strong> their variabilityincreases with the expected number <strong>of</strong> distinct original observations selected in <strong>bootstrap</strong>.Fitting <strong>an</strong> inverse power law curve to a series <strong>of</strong> repeated leave-one-out <strong>bootstrap</strong> valuesenables us to define a conservative estimate (not subject to downward bias) that providesa compromise between estimates with large variability <strong><strong>an</strong>d</strong> large upward bias. Inversepower law curve is a flexible way to model a learning process, <strong><strong>an</strong>d</strong> is quite typical indescribing machine learning, hum<strong>an</strong> <strong><strong>an</strong>d</strong> <strong>an</strong>imal learning behavior (Shrager et. al. [14]).Mukherjee et al. [15] studied sample size requirements in microarray classification usinga similar learning curve.4. COMPARISON OF METHODSIn this section, we compare the <strong>methods</strong> described in Sections 2 <strong><strong>an</strong>d</strong> 3 through simulationstudy <strong><strong>an</strong>d</strong> <strong>an</strong> application to a lymphoma dataset.4.1 Simulated DataSimilar simulated data sets are considered in Molinaro et al. [8]. For each simulateddataset, generate a sample <strong>of</strong> n patients, each with p genes (or features). Half <strong>of</strong> thepatients are in class 0 <strong><strong>an</strong>d</strong> half in class 1 divided according to disease status. Geneexpression levels are generated from a normal distribution with covari<strong>an</strong>ce10

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!