09.12.2012 Views

I__. - International Military Testing Association

I__. - International Military Testing Association

I__. - International Military Testing Association

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

The linear model indicates that the smallest amount of variance is<br />

accounted for by device features (.07). The combined other sources of<br />

variance: Instructor leniency, student ability, and task difficulty,<br />

(.21+.27+.42*.90) are predicted to mask out the variance due to the device<br />

features. Evaluators could also artificially change their ratings to<br />

reflect the impact of anticipated evaluation design changes. A<br />

reexamination of summary statistics would permit evaluators to assess the<br />

impact of hypothetical design modifications on the anticipated outcome of<br />

the device evaluation.<br />

DISCUSSION<br />

Using data from a simulation model, the training effectiveness analysis<br />

estimated that the one-trial difference between training under the visual<br />

plus motion condition and motion alone would not be statistically<br />

significant with a reasonable sample size (Ott, 1977). This outcome ofothe<br />

model was confirmed through analysis of actual field data (Evans, Scott &<br />

Pfeiffer, 1984). With this insight, from the model, the evaluator of a<br />

.device would know in advance that control of task difficulty, student<br />

ability, and instructor leniency in a field experiment would be necessary to<br />

increase statistical power. True training effects attributable to the<br />

device features are more likely to be revealed when extraneous errors are<br />

controlled. Cochran and Cox (1957) have presented a theoretical discussion<br />

of this problem. Instructors' rating variance, for example, may be<br />

controlled by utilizing a standardized method for identifying when the<br />

student has achieved mastery (Rankin & McDaniel, 1980). Some criterion<br />

measure other than instructors' ratings could also be employed. A specific<br />

example is automated performance measurement on the tactical range, which<br />

unfortunately is not widely available for scientific measurement of ai,rcraft<br />

in free flight. However, performance measurement is available in flight<br />

simulators. Computer-aided techniques for providing operator performance<br />

measures have been provided by Connelly, Bourne, Loental and Knoop (1974).<br />

coNausION<br />

This study shows that flight instructors who have knowledge of a<br />

training situation . but who are not necessarily proficient with the<br />

intricacies of research design and statistics can provide data useful for<br />

planning a field experiment (device evaluation). The programs described<br />

herein are "user-friendly" and resident in a portable microcomputer. Should<br />

the computer be unavailable, a questionnaire could be used (Appendix). The<br />

utility of this approach depends, in part, on asking the right questions for<br />

a particular training environment and in part on developing the responses to<br />

such questions into meaningful information. The model just described has<br />

provided that utility for the present situation. Additionally, this model<br />

may be easily adapted to other training problems involving expert ratings(see<br />

Pf'eiffer and Horey, 1988).<br />

197<br />

_

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!