15.11.2012 Views

Technical Report - International Military Testing Association

Technical Report - International Military Testing Association

Technical Report - International Military Testing Association

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

I--<br />

,<br />

c<br />

Now, an MOS Evaluation Test. rates examinees’ ebilitiec to perform the<br />

full scope of the duties of an HOS skill level, weighted to fit the<br />

objective3 and concepts of relative importance of the standards setting<br />

level of the Army pro&ream manager3 staff. In other words, the testo are<br />

biased in favor of the broadly-skilled examinee rather than one who is<br />

exceptionally skilled in a pnrt of his HOS skill level reqafremcnts.<br />

The test score indicates the exazinee’s relative standing among all<br />

examineee ansigned the MOS skill level. If peers and supervisors can<br />

adjust their ratings of exr=,inces to comyenaate for their personal<br />

bioaes and influences, if they thoroughly understood and accepted the<br />

objectives and important concepts of the standards setting level of the<br />

program manager’s staff, and if they could simultaneously and comprehensively<br />

evaluate the abilities of all personnel assigned the primory<br />

MOS skill level, they would probably rate the exnminee as the test rated<br />

him, if the examinee reacted to them as he reacted to the ceet during the<br />

evaluation period. Ohviot~sly, peers, auperviuore, and exainees interreact<br />

differently, differ in their opinions a8 to the relative importance<br />

of tasks, have varying degrees of understanding of the specific objective3<br />

of top level standard setter3, and have rarely, if ever, had the opportunity<br />

to evaluate the abilities of 311 persons assigned to any MOS skill level,<br />

let alone nimultaneously. The test3 do provide.a basis for predicting how<br />

well an examinee could do on the ; Jb when permitted to do the whole job<br />

rather than a oubspecialty and when motivated in the acne degree and<br />

direction aa he was at the time he took the test. A greet deal more information<br />

concerning his interpcreonal relationships and attitudes along<br />

with those of his peer3 and supcrvisor3, their rating attitudes 6J d<br />

abilities, and the dynamics of the group would bc required before one<br />

could predict Low an examinee would be regarded oy his peers with<br />

reasonable accuracy.<br />

A few item writing and technical publications writing groups have<br />

cooperatively used test question response data as a cue to whether technical<br />

material has been dietribuced to and understood by examinees. With<br />

few exceptions, the test question analyses developed by the Enlisted<br />

Evoluation Center reflect how all of the personnel in each MOS skill level<br />

responded to each test question. Items relating to equipment or doctrinal<br />

changes are carefully reviewed when only a small proportion of the CXaminecs<br />

answer the questions to determine whether the queetione or the<br />

cxaminees are deficient. Checks are also made to determine when guides<br />

and training materials were distributed to exeminees and whether these<br />

material6 require clarification or amplification.<br />

Many will envision other u3es of PiOS Evaluation Tests. In planning<br />

the uses, one must bear in mind and insure:<br />

a. The te6t content and objectives are compatible with each<br />

planned use.<br />

trolled,<br />

b.<br />

All pertinent variables are identified, evaluated, and con-<br />

____ __- .--.--- --<br />

146<br />

-.-_- _._. _ ..-.._. -<br />

- - -_ - -.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!