09.12.2012 Views

I__. - International Military Testing Association

I__. - International Military Testing Association

I__. - International Military Testing Association

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

The performance tests utilized product scoring wherever a product was produced<br />

as a complete or partial result of the performance. Where feasible, product<br />

scoring is desirable because, correctly administered, it can enhance<br />

reliability. The nine tests that utilized at least partial product scoring<br />

are identified as such in Table 1.<br />

In addition to the scoresheets, developers prepared equipment setup<br />

instructions, instructions to the examinees, and scoring instructions. The<br />

entire test was designed to be administered at a single station using either<br />

an actual or a simulated ship's radio shack. Although the 14 tests were<br />

independent, they were operationally interconnected so they fit logically and<br />

sequentially into the test situation and location.<br />

Written Tests<br />

Written tests were developed that corresponded to the 22 tasks covered<br />

in the hands-on tests. Three features characterized these tests:<br />

. The tests were performance or performance-based. Items were<br />

based on either performing the same steps required in the<br />

hands-on test or in answering a question of how a step is done.<br />

� The tests were founded on performance errors. To insure items<br />

were performance oriented, the causes of error in performance<br />

were identified. Error was identified as having four origins:<br />

The Radioman did not know where to perform (location), did not<br />

know when to perform (sequence), did not know what the product<br />

of correct performance was (recognition), or did not know how<br />

to perform (technique).<br />

� The tests provided likely behavioral alternatives. Incorrect<br />

alternatives were based on likely errors that were possible and<br />

do occur on the job. Incorrect alternatives also had to be<br />

wrong, not merely less desirable than the correct alternative.<br />

The development result was an 87 item test in a multiple choice format<br />

that was organized into 11 topical, functional task areas that generally<br />

corresponded to the 14 hands-on test areas. (Several of the hands-on test<br />

areas that needed to be treated separately for administrative and equipment<br />

set-up requirements were combined for the written test, and one written test<br />

task area did not survive validation.) These 11 written test areas were<br />

organized so they could be administered and analyzed independently.<br />

General Knowledae Test<br />

The third area of RM testing was a written general knowledge test. Like<br />

the written performance test, this was a multiple choice test and was based on<br />

the same tasks that generated the hands-on tests. The difference between the<br />

two written tests was that the written performance test was specifically<br />

designed to measure performance while the general knowledge test measured the<br />

application of knowledge to the task subject--which may not necessarily<br />

reflect performance. For example, the written performance test might describe<br />

a situation and ask what EMCOM condition should be imposed under those<br />

circumstances; the general knowledge test might ask what EMCOM is.<br />

530

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!