Technical Report - International Military Testing Association
Technical Report - International Military Testing Association
Technical Report - International Military Testing Association
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
I--<br />
,<br />
c<br />
Now, an MOS Evaluation Test. rates examinees’ ebilitiec to perform the<br />
full scope of the duties of an HOS skill level, weighted to fit the<br />
objective3 and concepts of relative importance of the standards setting<br />
level of the Army pro&ream manager3 staff. In other words, the testo are<br />
biased in favor of the broadly-skilled examinee rather than one who is<br />
exceptionally skilled in a pnrt of his HOS skill level reqafremcnts.<br />
The test score indicates the exazinee’s relative standing among all<br />
examineee ansigned the MOS skill level. If peers and supervisors can<br />
adjust their ratings of exr=,inces to comyenaate for their personal<br />
bioaes and influences, if they thoroughly understood and accepted the<br />
objectives and important concepts of the standards setting level of the<br />
program manager’s staff, and if they could simultaneously and comprehensively<br />
evaluate the abilities of all personnel assigned the primory<br />
MOS skill level, they would probably rate the exnminee as the test rated<br />
him, if the examinee reacted to them as he reacted to the ceet during the<br />
evaluation period. Ohviot~sly, peers, auperviuore, and exainees interreact<br />
differently, differ in their opinions a8 to the relative importance<br />
of tasks, have varying degrees of understanding of the specific objective3<br />
of top level standard setter3, and have rarely, if ever, had the opportunity<br />
to evaluate the abilities of 311 persons assigned to any MOS skill level,<br />
let alone nimultaneously. The test3 do provide.a basis for predicting how<br />
well an examinee could do on the ; Jb when permitted to do the whole job<br />
rather than a oubspecialty and when motivated in the acne degree and<br />
direction aa he was at the time he took the test. A greet deal more information<br />
concerning his interpcreonal relationships and attitudes along<br />
with those of his peer3 and supcrvisor3, their rating attitudes 6J d<br />
abilities, and the dynamics of the group would bc required before one<br />
could predict Low an examinee would be regarded oy his peers with<br />
reasonable accuracy.<br />
A few item writing and technical publications writing groups have<br />
cooperatively used test question response data as a cue to whether technical<br />
material has been dietribuced to and understood by examinees. With<br />
few exceptions, the test question analyses developed by the Enlisted<br />
Evoluation Center reflect how all of the personnel in each MOS skill level<br />
responded to each test question. Items relating to equipment or doctrinal<br />
changes are carefully reviewed when only a small proportion of the CXaminecs<br />
answer the questions to determine whether the queetione or the<br />
cxaminees are deficient. Checks are also made to determine when guides<br />
and training materials were distributed to exeminees and whether these<br />
material6 require clarification or amplification.<br />
Many will envision other u3es of PiOS Evaluation Tests. In planning<br />
the uses, one must bear in mind and insure:<br />
a. The te6t content and objectives are compatible with each<br />
planned use.<br />
trolled,<br />
b.<br />
All pertinent variables are identified, evaluated, and con-<br />
____ __- .--.--- --<br />
146<br />
-.-_- _._. _ ..-.._. -<br />
- - -_ - -.