15.11.2012 Views

Technical Report - International Military Testing Association

Technical Report - International Military Testing Association

Technical Report - International Military Testing Association

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

UNCLASSIFIED/UNLIMITED<br />

<strong>Technical</strong><br />

<strong>Report</strong><br />

distributed by<br />

Defense <strong>Technical</strong> Information Center ’<br />

DEFENSE LOGISTICS AGENCY<br />

Cameron Station � Alexandria, Virginia 22314<br />

UNCLASSIFIED/UNLIMI<br />

0<br />

( q


,<br />

a<br />

8X<br />

I<br />

, . ‘0.: I. ! ‘, . !.‘.’<br />

c L f ,, il ! !I 5 ti 0 u 5 t<br />

. :a , , , ‘, I, !. ;‘I! ..*<br />

,>. ,- ,-, - i ,.,. , .,! :;;r,:<br />

Fort himin Harrissfl<br />

Indiasapolis, hdima<br />

---. -s-e -_---._ - -.... -. -m-s-- U-Ryl’ ,:., j i .,. -<br />

-__ - --<br />

,


. r<br />

� �<br />

. .<br />

rnl.ITAkf ies?ING ASSOCIATION OFFICERS<br />

PRZSIDENT<br />

Colonel Jemar C. Doneghey<br />

Comaending Of iicer<br />

US Amy Enlicted Evalurtioo Center<br />

PRESIDENT-ELEtX<br />

Colonel John V. Patterron, Jr.<br />

Cowending Off fcer<br />

6570th Parronnel Rerearch bboratory, US Air Force<br />

STEERING colonnm<br />

Dr. Doneld L. Warno. Chairman<br />

US Army Enlirted Evaluetion Center<br />

Colonel June@ C. Donaghey, US Army<br />

Colonel John V. Patterron, Jr., US Air Force<br />

Captain Richard ?f. Hayes, US Nwy<br />

Ceptein J. P. Martin, US Coast Guard<br />

Lieutenant Colonel C. J. Chatroon, US Amy<br />

Lfeutenent Colonel A. S. Kneuf, US Air Force<br />

Ua)or Frank L. HcLaoethan, US Air Force<br />

Litutenent Caxaendcr Prances S. Turner, US Coast Csard<br />

Major Doneld L. Diamond, US Xarine Corp.<br />

Lieutenant 8. R. William, US Coast Guard<br />

Hre.‘Mabel 0. Brunner. US Air Force<br />

Mr. C. J. Hmeluro. US Navy<br />

SIXTH ANNUAL CONFEREHCE CCFBIITTEE<br />

Mr. Frank $3. Price, Cha.irman<br />

Captein Henry H. Bahner. Vice Chairman<br />

Hlljor Herril R. Owen<br />

Hr. William W. Wance<br />

Second Lieutenant Martin S. Brown<br />

SIXTH ANNUAL CONFERENCE PRCKRAM CCWlITTEE<br />

Hr. Claude P. Rrfdger, Chairman<br />

Hr. Dale R. Baker<br />

Hr. John S. Brcnd<br />

Hr. Arthur E. Hoffman<br />

** .<br />

The opfnions exprerred in the paperr prercnted in these proceeding8<br />

are those of the authorr end ase not to be construed as being official or<br />

in my vry representative of any of the US Armed Servicer.<br />

.<br />

F I


_ - --- _<br />

.._<br />

----.---- --- -- _____ I ._--.<br />

.<br />

-. -<br />

.<br />

. _ . - -. .- - ----- -<br />

_ . _ _ -. . .<br />

- ..___ - ._.- . .: . . .-___<br />

. ---.<br />

. . ’<br />

� �<br />

. * -<br />

.<br />

. . .<br />

.<br />

.<br />

---._.<br />

. . ,<br />

.. --.-... -__ _____.<br />

2;<br />

._<br />

--<br />

.<br />

.<br />

.<br />

:<br />

.<br />

.<br />

,<br />

c<br />

.<br />

.<br />

.<br />

. ,- ;<br />

.<br />

.<br />

.<br />

:<br />

-, “.,,


:a--<br />

�<br />

.<br />

. __ ._ _ .._ _ .___. - - - --_- __.. - -. ._ _ _._--- __.__ _ - ._-_. . -<br />

. .<br />

. .<br />

.<br />

.- __-yI----- ---.-- . . __I_ -..- --.. --- -___ --.-.---...-----


. .<br />

. .<br />

. . . .-.<br />

.<br />

.<br />

C8owal Syrposiuu It Tart Conotruction Proceduree,<br />

Williaa u. Wbnca, c!-4nnem . a . . . . . . . . . . . . * . . . . .<br />

*<br />

Tut Colutructicm 88 8 Subrprtea Within � �␛����� Approach<br />

to Training Technology, Iredora J. bt- . , . . . , . . . . ,<br />

Sumaq of Pragmatic Creativity in B+tiaatfon Conotructioa,<br />

John hedifotd . . . . . . . . . .<br />

Bvaluatloa of Notor’& il;, ‘&l&*E!: &&id; . . . . . . . . . .<br />

Performance Teot Construction, tred B. Horn .-. . . . . . . . . .<br />

08narbl Symporiur 1st Xntorprotrtionr and U8u of Bvoluaticm<br />

Ruulto, )Prr. Genwieve Schulter, Chairman<br />

One Ioterpretatfcm of the Major Goale of Specialty Knowledge<br />

Tutlng La the United Statu Air Force,<br />

. Stephen �� ����� � � � � � � � � � � � � � � � � � � � � � �<br />

� U8eI of )oS Bveluation Teat ReIult8, J, B. Rohrsfter . . . . . .<br />

caner01 Symporium XIX: Job Anolysir for Teat Devolopwnt<br />

Purpour , Prank I.?. Rico Jr. Chainam . . . . . . . . . . . . . .<br />

Neu Pmrpectivar in Job Anrlp8i.8, Joreph R. Harsh . . . . . . . .<br />

)soS Ev8lurtion Teat Outline Sadnat, Curt18 D. KcBrldo . . . . .<br />

Group Dfrcurrion I: Ita, Writing Procedurer for Incremiag<br />

Validity, I. J, Newnun, Chafman . . . . . . . . . . . . . . . .<br />

Group DircuorFua 11: Non-Bnpirlcrl Velidbticm of Tart Item,<br />

J. L. Pinucans, Chaima~ . . . . . . . . . . . . . . . . . . . .<br />

Group Diacurrfon III: Te8t and Itr Revf~im Techniquer,<br />

* J. g. Partington, Chainsan . . . . . . . . . . . . . . � . . . ,<br />

Peprrr An8were t o Camon Criticimu o f Tolta,<br />

, t P~kR.RIceJr........................<br />

1<br />

Camnlttm <strong>Report</strong> I: Sturfng Comdttee . . . . . . . . . ,, � . . .<br />

Canmittm Rqmrt ZIr fca Writer8 Aptitude Tut Devolopaent<br />

onmlttu � � � � � � � � � � � � � � � � � � � � � � ��� � � �<br />

Roster of Confereer . e . . . . . . . . . . . . . . . . . u . . . . 184<br />

._~__ - . ^ ._ .-. .<br />

.<br />

.<br />

_ --‘T’--- -<br />

I<br />

e<br />

----<br />

.<br />

.<br />

.<br />

,rcc<br />

. .<br />

.’ -<br />

.<br />

.<br />

:<br />

.<br />

.<br />

121<br />

122<br />

126<br />

127<br />

132<br />

136<br />

141<br />

148<br />

149<br />

156<br />

160<br />

161<br />

171<br />

173<br />

181<br />

182<br />

P<br />

4 Y


.<br />

.<br />

.<br />

_ _- -_ _.. ..--._... -.. .-...__- --_- __ ,_ __ -<br />

.<br />

“1<br />

. I’<br />

. .<br />

__,._<br />

- F&r.. L...A.-.-.----..<br />

w-h_- _.__ _I_____._. I _.____..-me.- --_<br />

_______ _--------- --..- . .-. 5<br />

Foreword<br />

P<br />

Unharaldod at its inception Ln October 1959, lhs Xilituy Tsrting<br />

A88ocLatioo h88 dwelopul into a povorful influence in military tutlug.<br />

The purp03s of tha haadei.atiou ir to bring together annually rrpresontatiVM<br />

froo. the V8rioU8 ~m8d 88rvic08 to difKu88 and exchaqo idaas<br />

. WUCOYUtKlg job PrOfiCtOUCy WahatiOU Of enlirtad p8r8CWl81. nU8, one<br />

of the more important 8nd productive -pact8 of ‘:he annual couforanco<br />

i8 8 WOting of mind8. Confermcer h&v8 bean hold rithin 8 liberal<br />

. .A fr-k of camftter mestingr. While the liberty of ccmuitter dircu88ionr<br />

tur provided opportunity for crsrtive thinking and productive<br />

dircurrion. the rtructura ~sca888ry for exchanging &8 much informtion<br />

88 porrlblo rithin A thort ti.m wa8 lacking. h progrrn for the Sixth<br />

Annual Coafumca ~88 developed vithin uhich auimm opportunity would<br />

oxirt for the prerant~tkm of rorurch rwultr and operating lnfornutioo<br />

through tha prerentation of preprred paper8 and for the exchmgm of<br />

id-8 and crutive thinking through group dircusrion8. fh8 eonfsrsnca<br />

wa8 divided into two theoretical rad tw geueral 8ympori8. P8plrr 88<br />

prorented in the 8yPpo6ia 8nd report8 of group di8curriou8 are roprod&Iced<br />

fn thooe procauiiagr. It ir mticipat&d that there proc8ediagr<br />

of tha Sixth Annual Conference and the procoedingr of future couforencos<br />

~111 becau8 a 8ource of infornution on poychologicrl teoting in the<br />

ni1itaX-y 88tablirhment.<br />

. r > *, ‘. ,k ‘*> y,,. . .‘J 1-. i<br />

‘- The Sixth hnnu8l~bnfsre?e8j;]a8 hdld et the US Amy Enli8tad<br />

Evolurtion Center, Fort B8njtmin F~T~~OTI, Indiana. Tha c-ding<br />

officer and staff of the gnliotrd Eva?uatFoa C@nt8r -.irh to uxpreor<br />

8pprcciation to a:1 COnfOr for thoir rupport and ccanplrts cooperation.<br />

Tha 8ucce88 of thr Sixth Annual Conference wa8 dua to thr attitude and<br />

offortr of the participu\rr.<br />

iii<br />

.<br />

4<br />

,


__. . . -- .;- * . -_<br />

.<br />

_... - - - _<br />

+<br />

. .-._<br />

.<br />

. . w<br />

1) I. .-c -. - _ . . - _. ..--z--.-m - _..--._. -_- -_- -. _ .._ _..--- --.-..-w .Mrm. _<br />

I<br />

I<br />

1<br />

KCLXTARY TKSTXRC AssocIATxm<br />

Sixth Annual ~oafsrencr<br />

Colonrl Jm.rs C. DonaghcF<br />

Prwiding<br />

1


.-. c 1. - - . . __.<br />

.<br />

_ _ _ _ . -- - --__.- -_ - . _- --.- -.. .A---- - .--.. _.-___ - _--.---.-.<br />

.- - - -<br />

, .<br />

.<br />

.<br />

_ _. _ _ ._ I<br />

!<br />

3<br />

i<br />

~-RUB* ----_ _--<br />

.<br />

.<br />

~--~~----.A-‘-- .--.--.---- - - - - -<br />

I’<br />

a<br />

a - . - - - - . - - - .-<br />

!<br />

I !<br />

I<br />

. .-<br />

.<br />

\<br />

Rtnrrhr to<br />

Wilttrry Tortkg As8oclatloa Conference<br />

Ilepui~~O~~~l<br />

QMlce of Per-1 Operation6<br />

Deptubent of the Amy<br />

Tllmk m cool JJQw!%Y,<br />

I mnt to beginby exprecalng pry appreciation ror thle agporttmity<br />

Ican'tstsadberstoda;yandtdUtoyous6atastprryclhologletor<br />

expertlnpsychologlcal meaeuremEtibec6ll6eI6mIlot. IamzLnandIhera<br />

bMn for scoSx’6l par6 Mlrigned to pofdticm Where 1 hwe U66d the<br />

lxfonaetlon that you people product. In w prevlou~ ssel@mt, I have<br />

lhmedto eppmclatethe value of MIundtcrrt ma-t in h%lpfrq<br />

aDyon6 86eigmd a task and trybg t0 Eake the bO6t pO66ibh we Or<br />

per6omelzvoursesto ccmplata the task.<br />

9%~ cvalustloo of job proriLicncy orrem 6 challenge that 16 not<br />

easily overcam. bthfE&eyof eXpert6 and cadvanCedtCChtlO~~, it 16<br />

t%lf+ffcult to expect an accurak evahatlon of ld.ltidual potential<br />

accazplishcd with peper ard pencil. I 6.m rirmly cormfnced, htmmr,<br />

l&atycu,havetoalnrgedegree eccaupliehedthi6end. Just a6manag6m6ntbytheru.bdtbeth~6ibhssbeenre~cedby~<br />

with a<br />

633d8 ld.0, 80 too ha6 OYtdu6tion by fntUftiOXi been nphCad by 6ClOXltifiC<br />

eralMt10n. (3exztm, you have pa66ed the point or no zxturn ror<br />

ob;Jtctlm evalutlmof job mastcry. We in the Anuy are rlmly conW<br />

th6tOUl-Fnflstcd~~t~OXlcSyat~16 &FetO St&y. IbO~fB~that OUl'<br />

EnU6tedgv~tloaSy6temletheamo6t eclentlflcalJyedv6mxitoolwith<br />

vhichv6haoatoQtebeenablatoprwidethec~ r and hle per-<br />

6omel6tafY.<br />

Ursfoz.-tumtely,Im6t ln6ux-t thstws have notbeun able to instill<br />

in our cnnrmRndcr6 a degree of confidence in the sy&em which ym vcnild<br />

lfka to achlow. 6iovthatcoloabl~~yardhi8paoplsha~devaloped<br />

instmmentathaten~thmliable andvaluble lnmewurlngthe indivf.dud<br />

job proclrlsncy, my miesicm frcm Eeadquartere, Dcpartpnn~ or th6<br />

Amy16 tohelpthe6epeople developtb6 undemtmdlng andcoxLfldence of<br />

cur lbry c-ers. If I mkn no other point todfiy, I reel- to<br />

1<br />

.


,<br />

.<br />

M<br />

‘, .<br />

i<br />

i<br />

,-<br />

‘.<br />

.<br />

toll you th6t th. grrotert short-coalng in poychologlcal test wuraomt<br />

f&y i8 th6 16ck of undsratmding of proficlrncy 6ValMtfOft by the V6w<br />

p6OplU1 thrt muot U06 this tool. without 6 knouledga of th6 teehiqueo<br />

fnvolvsd, memagms6nt i6 fr6qu6ntly 810~ to 6ppreCiat6 th6 Cool th6t GUI<br />

b6 provldod by thlo 86rvfc6. Th6 tnliot6d Evelu6tion C6ntor h6r6 ho8<br />

taught 016 much of wh6t I knou about poychologicrl toot m6aourement with<br />

+6p6r and pencil. f don’t pr6tend to know th6 d6t6ilo--that io not w<br />

job-but I do beliov6 w6 h6V6 oaa6thing bore that m can n6v6r 6g6i!¶ do<br />

without 6nd thr largest tuk reaoining Lo to ukr those uho arot u66 thio<br />

infOrPrartiOn avare Of it0 Vo1U6.<br />

Enlioted rvoluatlon In th6 Am7 go60 b6ck long bofOr6 th6 curr6nt<br />

oyot6a. I can racall th6 Army’8 Carwr Cuid6nc6 PIan of 1940 vYIlch W6o<br />

6n 8ttempt to ~06 testing 0~0~66 to muoure prOewtfon pot6ntial md<br />

d6t6nain6 thr ultlm6te promotion, prior to World War II. Going back 6v6n<br />

prior to World War II, th6 lndlviduol br6nch68 of th6 Amy devrlop6d t6oto<br />

to 6ld thmo in the promotion of noncanioofon6d offlcrro. The � ␛���� w<br />

h8v6 in the Amy today go68 fro b6yond thoo6 iaftf61 6nduvoro. W6 ~06<br />

our 6nlioted 6r6lrution to doterain rinimum qu6lificrtfon in th6 rilitmy<br />

Op6Cidit7, to d6t6min6 th6 6w6rd Of prOfiCi6nCy popbnto for OUp6riOr<br />

quallfic&tion, 6nd u6 ~06 anlloted 6volu6tioa 48 8n fndfcatlm of potsnti61<br />

for th6 promotion of our enlioted pwoonn6l. I 6o1 our6 moot of you 6re<br />

ware of th606 rpplic6tiono for our � yotao of mliot6d 6V6lUotiOn. w6<br />

6100 u6e enlloted wolu6tioe for gr6da det6rmiMtfon for psroom6l r6vrrtlng<br />

io UI enlfotrd ot6tue. It lo not gmamlly kncml that W6 eloo<br />

Uo6 the 6UliOt6d WoluetiOU oCOr68 for 6OOi~t o618CtiOn purpOo60 at<br />

department81 16~61.<br />

In ths Enliot6d Poroonnol Dlrectoret6, ths 6ooignr.?nt of our<br />

mOt6r � ergunto, firot � �������� and 06rg68nto mojor io wAd6 ~1317<br />

after 6 couolder6tion of th6 6VolUotiOIl 8cor6. WO h6vo eotablirhed<br />

c6rt6ln minimum 6volu6tlon 8cor68 for 688igmrunt to hlgh priorIt+ and<br />

critical pooltlono. W6 hovr 166rned thr value of 6nliot6d evrlu6tion<br />

� ���� ����� ���������� ����� �������� �������� fOt ������� ������ for<br />

obvlouo ruoono, w vant uut b6rt ~60~16.<br />

Thr ~06 of our svolu6tion q8tm in th6 Amy h66 grown by leopo 6nd<br />

boundo. W6 hove r6centlp lnltiot6d 6 progrmu for 6voluating 8 roldi6r<br />

Ln okillo oth6r th6n hi8 prinmry. W6 er6 tqiug t0 68t8blioh 6 valid<br />

oklll 16~61 in wr6 th6n on6 op6ci6lty. 6n6bling uo to know fully tho<br />

pot6ntial of our emlioted p6roonnrl. W6 lik6 to f661 that thla<br />

axpuloion of our ev6luotion c6pebiliti60 will 6nabl6 us to Lee thr<br />

Who16 ooldl6r, r6ther thon juot the primary � killo of thr ooldi6r. Thuo,<br />

6v&luotim of ~rcottdarg okillr will 6n6bl6 uo to wrov6 the uoo of our<br />

omnpov6r r6oourcao. Th6 old FrOblu i8 that c6rt6in ok1118 ato n6odod<br />

in grut6r proportioa owrouo than fn th6 Unit6d St6too. If ve only<br />

2<br />

.<br />

.<br />

-A___ ._..--<br />

.<br />

__...--.---<br />

.<br />

*<br />

---<br />

.<br />

I<br />

.<br />

.<br />

.<br />

I . .<br />

. . . 2 . .<br />

’ .<br />

*.<br />

- ‘.<br />

,. ., i<br />

.I b<br />

1.


(i<br />

.<br />

* -. -._ _<br />

. .<br />

.<br />

.<br />

have one @kill identified for the roidier and tbst skill 18 one uhich<br />

lm need priorrily overrear, that roldiar doesn’t get hi8 fair ohare of<br />

tima in the Waited St8ter. Utmn’ two or more rkillr are idatiffed,<br />

more equitable dirtribution of ovormea duty ir posrible.<br />

hnothor uso of the Eelirted tvaluatiq System ie the waluation of’<br />

the Army Rererve uld Natiou81 Guard enlistrd personnel. Rvaluation of<br />

there people vi11 ‘aid ua in determining mbilitation pxentlal a8 wall<br />

&8 providing an edditiocml Incentive for masuberr of the Rcmrvr mad<br />

Mational rwrd to maintain tholr milltarp rklllo. We are extremely<br />

proud that our Enll8ted Evaluation Sy8fem has been axtandrd to our<br />

enlisted Reoerve carpoaaetr. We consider this � otep in the furtheronce<br />

of tho ona Army concept of teamwork betworn tba Army compomntr.<br />

tvaluating rrcondary aed additional rkillr and evaluating perrmoel<br />

of thr Rsaem compoearta hu b88.e made porribla by tha u8a of tho Hcil<br />

315 Sy8tm, vhich w na uaa to handle tha ta8t data here la tha Enlisted<br />

tvalwtioo Cent*r. rtre uao of � utaatlc data procraring bar enrblad UI<br />

to look fonnrd to future application8 for our enlisted ev8luation. Flora<br />

at thir iertallatfon next month, rrq-creotatfvar of the Continental Army<br />

C-d, my offica, and talioted Evaluation Center ~111 meat to dircurr<br />

the potoetlel fur providieg a training evaluation report to our ccnmmderr<br />

fras thr ccmpaay love1 to the field Amy. If the proreguirlte rkillr to<br />

8 glvm Job can bo Individually evaluated categorically, thy CM aleo<br />

bo evaluated end ourmarired bp the unit or by the’ cpaciallty.. %lo vould<br />

enable tho unit comsander to deternina area8 vhlch require mupharf8 in<br />

hir training progrm, em ~811 cd helpins u8 doviro ochool program which<br />

vi11 insure tho � daquata troinfng of pcrrroenel in their military rpecialty.<br />

We na dlvide.our evaluation of enlisted porronnrl into flvs to ten araa8<br />

appropriate to thr given specialty. It ir there areu vhtch we Lateed<br />

to analpre for -raining defic.iencfrr. It 18 anvirionad that all our<br />

canandorr vi11 be able to use the informetioe which ~8 provida. I wka<br />

this point hero becaure I want you to underotaed that we feel that the<br />

product of our enlirted evaluation rfll be ured by our opcretlom people<br />

and our training people. tnlirted Evaluation can and will rupport the<br />

vary hurt of our Army mir~ion of rudinerr to ������ � � uie lo limited<br />

or general ground combat.<br />

I vant to conclude on thir note. The contribution8 of objectivr<br />

rvalwtioa of indlviduel Job proficiency to the mi88ion of the maed<br />

rrrvicer of the United State8 are inmururable. I an convfecod, aed in<br />

fact I kwu, that more and wre of our ronior officer8 aro rocogeisieg<br />

&fly the catrlbution 0.’ proficiency evaluation to the task of tumagrmeat.<br />

A8 rout cyatrn becuner swra rophfrtlcated mnd you ruch for<br />

perfection in ths raliabiltty cmd validity of your teOC8, you Wu8t<br />

novor forgot &at vlthout the confidence of mnag8mut, your product<br />

doer little good. The ao8t valid evaluation � ecomplirhor nothing vhbe<br />

.‘<br />

. :<br />

I<br />

:“i f<br />

, .<br />

i<br />

1<br />

.<br />

I<br />

. , I i<br />

-. ---.<br />

7


it falls on the deaf eara of mt. Tberhalldngebef~youis<br />

lmm t?.iaD mm?4 illpmhg your f%al.uation, it. ia 6ou the uee of<br />

objactitre ovahlation for the many prpo6oe which at can 6ome. MO<br />

know that encb of you ViU contribute, In tb3 cow&o of this conferax,<br />

things that will help u8 llzgxxm our Imisted Eva3,uation SyrJtela. l&3<br />

hop0 that in your tima hero with the Fhluation Centor, we can offer<br />

you ~csaethl.ng to tab back to your jobs.<br />

Ithankyou.<br />

.<br />

Y


.--<br />

A-<br />

.<br />

__ .-._ -.-.. - -‘--- .- -.- :<br />

. _ ..-- - . ^-^ _ _ . . _ __.. - .._ - _. __. _ ~.<br />

.<br />

Remarks to<br />

Ifilitaq <strong>Testing</strong> and Aesociation Conference<br />

X&JOB DC&AID L. DIAZUMD<br />

Officer in Charge, Teat Sectioe<br />

US l&rim Corpe Inetitute<br />

Good afternoon Iadlee and Gentlesten. I me lookiag fonrard to<br />

having the opportunity of mingling with you gsntlaxn who are euparte<br />

in tha teeting fiald am I have very limited axpariancs and kawladge<br />

of thie aubfect. I vi11 briefly axplain the Marine Corps Teeting<br />

Syeteu by firet etatiag the laieeion of the Harine Corps Inrtitute.<br />

Our mieOioa lo to prmtids correepondonco coureau in baoic zailitary<br />

eubjacte to aelisted liarinee and :o prepare and procore sxaminatlone<br />

es directed by the Cawuedant of thr Marine Corpe. We have a third<br />

miesion whfch oftan makae It difficult to accaaplieh the first twoto<br />

provide ceranonial troop0 to the marine barracks for participation<br />

in weakly parades during the � wser month0 and other carcmumial c-ituente<br />

throughout the year such � e military fueerale, etate arrivale of<br />

foreign dignitariee, and other White Xouaa functione. To gat back to<br />

‘tha teeting bueineee,<br />

the teete we prepare and adaiaieter ure not pm-<br />

foruance evaluation teeto but are ueed to maaeure the szminee’e<br />

knwlodge of general military eubjtcte. WC prepare and � dminietor four<br />

groups of teete.<br />

The first g&p IO the Officerr’e Adminietrative Subjacte Exaeiinetioe<br />

which is aduiniotered twice a year to lieutenants and captaine. The purpose<br />

of thie sxarsinatfon ie to motivate company grade officore to fazeiliariee’<br />

thameelvee vlth variouo adminietrative type subjecte with vhich<br />

they will becoma involved throughout thelr’careere. The subject? are not<br />

tactical uor technical in naturo, but cover ouch � raae ae law and Lagal<br />

Mattere, Supply +&nagaeant, Plnanclal Haeagamant, Pereonnal Administration,<br />

Caneral Adminietratiou, and Organitation, Camand and C-d Relationehipo.<br />

The officer lo � dministared a teet ae a lieutenant and agafn’ ae a captain.<br />

He muet pare tha test or retaka it the subeaquaet yaar until ha doae pare<br />

it. To data, at laaet, the reeulte of thie teet do not reflect upon<br />

his fitnese for prcaeotiou � e the promotion board doer not eea the teet I<br />

raeulte. However, if the officer IO a repeated failure, hie comaeding<br />

$<br />

officer met rarka a<br />

c oeeeemt to thie effect in hir fitneee report. At<br />

this point, the toot reeulte will reflect upon hie fitneee for promotion<br />

as the praeotion board views all fitness reporte.<br />

The toot itema are prepared by a board of officere ealected from<br />

Haadquartere, Marine Corps. They are put into teet ����� � duinietered,<br />

� corad, md reportad upon by Xarlne Corpe Inetltute.<br />

The � acond group ie tha Central Xilitary Subject0 Teet which ie<br />

edeieietered three tine8 a year to corporalo, � ergaante, and � taff<br />

sergeants, both rrgulare and raearvee. This test ie ueed to deteneine<br />

thair eligibility for praaotion and muet be paeeed before they � ra<br />

coneiderad by the Promotion Board.<br />

.<br />

.<br />

5<br />

.<br />

,<br />

i<br />

.J<br />

.!<br />

.i


*<br />

1.2<br />

P :<br />

I ,<br />

. ’<br />

. c<br />

,<br />

. .-.<br />

. .<br />

-. .<br />

. . L<br />

-.--- ------<br />

_ _ ..-.~ _- ._ * _ ._- _.._. __- ____.__-.__ .- ____..__ - _-_-- --e-v-._<br />

In the Marine Carp8 we expect all !4arinsr, rogardloclr of their job,<br />

to be potential cozabat mariner. Thereforo, thio toet covora thoco oubject8<br />

vhich aro conridored to be ind18pen8able for wrines who may ba<br />

raquirad to perform duty ,in a ccxnbot aree. There rubjecta include Squad<br />

and Platoon Tactic8, Cuerilla Warfare, Scouting and Patrolling, NBC<br />

DOfenle, and Pir8t Aid, to name a faw.<br />

Technicnl proficiency within occupational field8 are deteminod by<br />

meena of fitnrlr report8 and proficiancy mark8 rulnaitted by the mmrino’o<br />

commanding officer on a periodic barris --but aot 1ePa than once every 6<br />

montho.<br />

Tho third group of tert8 ir the General <strong>Military</strong> Subject8 Proficiency<br />

EvalUation Tact prepared once a year and prcvided to unit C-dOr8<br />

in the field. Tboy a&inirter, score the tndt with tamplateo provld8d by<br />

MCI, and evaluate the rerultr with rt8tictical forma and inrtructionr provided<br />

by XX. Th8 pUrpO8e of thir tcot 18 to enable unit camandorr to<br />

avaluato their own Unit and 1ndividurl training progmnr urd epot the<br />

WUkIIe88e8 in variour rubject 8rOa8. The 8Ubj8Ct8 covered by thio tzot<br />

include thO8e previourly mentioned in the preceding tort plus: Clore<br />

Order Drill, Hilitary Courteey and Diocipline, Wap Reading, Comnunications,<br />

Milltarp Training, Danertic Dirturbancer, and Technique of In8tnxction.<br />

A differ8nt te8t lo prep8red for each of the three rank groupr-one<br />

for private-lance corporal, ona for corpxalr and rergeante, and one for<br />

otaff NCO8. The t88t range8 fran 1‘75 itcrrmr to 250 IteM whsru8 our<br />

othor tertr era limited to 100 itsmr.<br />

The fourth group 18 the Inopector Goneral TO8t ured by the XC te8m<br />

during their annual in8pectfon tour8 of 811 tlrrine COtp8 unit8. Th88a<br />

tcrts are ured to ev&luata the U’Ait training progrm � l8o. There te8t8<br />

� ro prepared to te8t the oame rank group8 a8 the previoU8 testr. Th8 Ic<br />

team � dmini8ter8 the te8t8, rnril~ the rerultr to MCI for Ilcocing, evaluating,<br />

end roporting rho ra8ultr back to the unit 8nd the IG teas.<br />

t!CI 18 currently prqaring rpecificationo for a computer syrtem to<br />

br inrtalled, hopefully, tow8rd the end of fircal yur 1366. Thio 8yrtan<br />

will be a boon to our curtent ta8k of evaluating corrarpondence cour8e8<br />

and the variour te8t8 which ve adminirter and Icoro. At thir point I<br />

wuld like to uprerr en appul to tho88 of you who h8ve computer inotrllationr<br />

rupporting either correrpcmdence Cour888 or your torting progrm-pluse<br />

cont8ct me at your earliert convenience end oxplain your ryrtemr<br />

end operatlonr. I wuld like to carry back your idea8 and rmthodr to<br />

incorporate into our propooed 8yrtem rpecificationr.<br />

In cloring, I an looking fonmrd to gaining a lot of valuable<br />

informtim to h8lp WI end the U8rino COrp8 in our future tortfng<br />

prograar .<br />

Thmk you:<br />

.<br />

6<br />

_.. _ _ . . . _-_<br />

. _1 -- --..<br />

.<br />

e<br />

.<br />

. . .<br />

. .<br />

.<br />

.<br />

J<br />

.


:<br />

i<br />

f<br />

I<br />

.<br />

.<br />

_.. _ ._ ” .-. -. ._ _ ---_ _-.-.. _- __._. _ c.-----_-.__- ___. - ..--.-.- -. .<br />

.<br />

.<br />

. .<br />

‘V-u--. -- - ___ ._. - -....- - ---- d- .__--.I_.._--- -- .-.... - .<br />

. .-. __-_._ -----.--<br />

.’<br />

.pai%rko to<br />

Ithasbeen auggeeteplthatazyrexuarks shouldconcerntheEXEUF6D<br />

uxlg in armtd-6ercTices examinations. SincamanyoftheMhlmmbershere<br />

todapnaybeas~~thateetingBamcasIam,lt~~proper<br />

that~lookboth~thepa~tandppneentbefo~ews~toZook<br />

to the futuzw. l!heresretwoquestlonsuhichbotholdaMnevmmbers<br />

mustta!u3tipssto considerbeforeb attempt toevsluatethe futursr<br />

1. Xaveuebulltourhcmeuftetrtlng6ub6tanti~ Isour<br />

rauudati0n alAd?<br />

2. Canw safe look bta the f’uture? Are w mare of our<br />

----?I<br />

~t~tblr? Are ve aoun in our prarent performance?<br />

!L%e snmerstoboth qumtloas becntobe positive. The pest rep.rts<br />

of our organization indicate that we have been united in our actitity.<br />

t&3 have three c~ ereas for self-pfaiec:<br />

1. We ten bmrt af our conwte attention to the vast mount<br />

or detail favolved in lndldbg summful testing witere.<br />

2, our c~acernwiththeproblcaof creatingbettertests conthluel<br />

to be urupmittlng.<br />

7<br />

‘_<br />

.<br />

.<br />

.<br />

.-


-<br />

,<br />

/ .<br />

,<br />

_ _ . _ . . . * .__.. -<br />

.<br />

m<br />

\<br />

. .<br />

.<br />

. -<br />

-<br />

__..__. _-- .-- ,-- - -.<br />

acccaupllshmint of our original alms has beccowof r&nor iapxtance to<br />

the und.reeoEd of expfmsicm that haa oecttred. In tha early fbqe, wa<br />

couldproject anicbalthntwsona ofhorizontalclai-. #s could<br />

veil lmaghm and work for impronmaats inquautycup?-oductionard<br />

tridnlng, an Increase in the rnrnber of tecte writton (euui at the tiESb,<br />

incraaecsiapor~onaeltocerryouttheloed)arrtaninmaeein~<br />

preatlge aftha indivlhal ixnters. Host of.tblshasbet~d.one--althm&<br />

it wsn% until the 1363~coderonce that WJ 8brted to mrk cm ~1~j3<br />

dellbarata~a~etomakethevalussoiourrrorklnww~.<br />

TIna has proved that the acccesplishmmt of our ndealon ha8 bean a<br />

lrtcp not 6n end. khavcincre55edladepth asrrrallaal&erally. Ve,<br />

theMIIA,havedcocloyaddbvicos,Prans,aJsdtecfmicZuasthatholdcballeDglae<br />

possibllltlee.<br />

lb havereachtdthc pointvhereve canexploretbeuut.er~cesof<br />

tQstl.ng: ua sro phyaicauy pYTqm?d. UO5X9OIlSOUdgD.d9~aUr<br />

takeoff. E?utarowpyd.10lnglc8llymsdy? Areyoureacl# Dow<br />

relax In cm tmethly nmning operations, or do vu met the challenge<br />

of what can ve ~DP pfy aoureeb regm-dlng the M.I!A'e poeoible &FIIVA fn<br />

exploxd$fi~eewtest~corrmoe are the reprtsoftbe pmwl0us cunrem-8.<br />

Lf1maytakethlsllberty,itwmulc? seemrraa them xvports<br />

that In ears aretm theHl!Ah.aabeentoo con-tith -ata Curicons<br />

to recofi~ Ga that the space beymd thea Is real~&%&i%<br />

UnllmLted." At the lWmlExminlngCe&erwe b.ave reachedthe poLt<br />

Mm-13 the order&q of emunlmtionsbytbe Fleet, apcessthathithmto<br />

took 742 man yea-8 per exmlnatloapcriod, cmnowbe acccmpllahed<br />

automatically. ~1ongorbwoeachunft~~tospcndthatiEsIl~tinS<br />

ztect?searyexeuldnatlons. !rha IEW autaaation nov coi7rplQtQs tho onttre<br />

mund o.? the examlnatlon process, In rddltiun, we have reac.hQd tha<br />

point whem itidividual =tura~ 5n3 6tia8diccilly r0m-c-l-<br />

PAKt system 80 thatallpersomelrecorda iwe ~~~@.~telyUg-to-clnti<br />

d.lwst 68 Bean 58 the individual 8COm 16 knUW!L %!hib affords m<br />

0fYicicnt distribut2oaofenllfstedperm.melthr~the~.<br />

In pas-t M?A zncotiags, acc~to the recoxYla* w ham been concerned<br />

tith the "nuts and bolt&" OdLy In 1963, vlzaa wo att-apted to<br />

glintbe recogxlitlonorDm,didwl dupartrrcmpurQlyut1lltarlan<br />

8834Xtt3. Wo have gdaed great twhnics3. itduty. RJW M mu8t CODcentrate<br />

on valulcr uf tmtlng, We rru3t roach out ml oxplor8 new<br />

.-. _-._ .______ . - - ---._-..--_.---_P_--__-<br />

.<br />

. . . .<br />

. . .<br />

8<br />

I<br />

.<br />

__ . _-. .-. . -. . -<br />

.<br />

. .<br />

f<br />

1I<br />

1<br />

f<br />

.<br />

Y


-.<br />

/<br />

,’<br />

i<br />

_. . . _. __ _. __. .._ _ ___ -. __- .-- -_ - ._ __-_ . _-.-___ -___ .__-..<br />

,<br />

. .<br />

. . . . . . --..-_ _ _._<br />

.<br />

. _ .__- .._. --__--. -- - -_---. . - . i ._..... -<br />

_. _--- _--. ------ .-<br />

.<br />

\r<br />

*<br />

Perhaps, BI a group we have teMod to remln Btatlc. Ia this<br />

becaueevofeelthataurinterastinaduc8tloDdhprsgressls t3ub60rrri0at<br />

tommber&lpinourtterrlaet Allafourllttaeaxwdevotedto~asas<br />

sspact of ta. )ie 6houUtharcfarercdiza that education ie one<br />

0fth8~6tp0wtr0~t3int.bk3~23f30f~.<br />

Slnc8 thetop~llftlce in the ileldaftwtlngam gathsmdbere,<br />

l!Betln&oxlacarmaDa gzpnmd ai f.xrterest, ham!t WeI tha ctblc8.l right to<br />

rtmoln etatic--tohats ouroretlrg6mmdy~nicelyorganized commutlan?<br />

1 umpasltlvem allbava a etrongtdth lnthu lnUvldualpotmrtlLl<br />

of the HEA. WebelLeve Rtroaglythatlt canbetcsmths (zEnAmmm<br />

IXTESTIRIwer&mm. Wsbulient&tveereprqxmxlardab&to<br />

arp- unknovnrields. Wsbc13evathst0u.rarp10r8tlsnati~bs mmad.<br />

I& us xmf, at thlr conPcxwlce, use aar CaUctlvd craatira lnBgiMt1on<br />

to aseart omwelwa<br />

,<br />

, . .<br />

.<br />

I<br />

. .<br />

I<br />

i<br />

4 ; i


* .<br />

_ . . _.-* --. .___ . _ _-. .- - .-.-<br />

.<br />

.c -<br />

.<br />

R8mA.B to<br />

KIlIta.ry Tertln~ Asaocintlon Confetanci<br />

LiQUtMQit CO1OCi81 8. 8. &UDf<br />

f o r<br />

COLQSSL JoBlD V. PXRZRSoB1, JR.<br />

ccQal8xldIn~ Off Icer<br />

657Otb Peroomrl ~srurcb laboortory<br />

us Air Ports<br />

Tuo Itcu of probable I~torart to the eoafrrear of tbr Wlltory<br />

Tutlng AssocI8tIw concwnIlq the Air torte SpecIalt7 KnwlQdge T*rtI!lg<br />

progmm are:<br />

(1) The clorar relatIon8bip developing betmm Air Force<br />

trainfn~ ahd Specialty Knowledge TutFag, ukd<br />

(i) ’ Rramt developmnt In tmhnlquoo of job Qnal78Ir.<br />

Although both of there, tolpiC8 wIl1 be the 8ubjQetr of tiditidual<br />

pruantatiour during tka cesfareaco, ‘L would 1IkQ to-brtefly Iatroduca<br />

then.<br />

Apptcolhatoly a ya8r aa0 tha Air Force 8dopted what 18 CilllQd (c<br />

&Sal-chmxml approach to on-the-job treinirq. tfadar tbir concept en<br />

airn~n ir prwided with 8 rorlas of Cue&r lMvQlcqsen?mt C~U~~QS ubich<br />

(l;r@ i.ntsndsd to cwer thQ broad fUd8WXltal8, theory, :md principles<br />

nhlcb ho Is raqoirQd to kaw to progrors or bo upgrQd@d ia hi8 QpaeIalty.<br />

‘LbO8a c0urs.u are dQvQ1op.d by Air Troininf Comaamd t8ChZliCd ldtQS8<br />

and prfntod and dI8trIbuted by thQ ltptenslon Cwrres Inrtltutr o f thr<br />

Air thlvorrity. Tb~y are Intenbad to be rQlf-study type cmr8e8 ubIcb<br />

the 8iSmM COLBplQtU 8 8 OZL ordinary COKSe8pond8W~ COUtQB.<br />

ibe other channel of tho &Ql-churorl CO6CQpt I8 BCtrQ COnCQtnCd<br />

with the 8wifiC tS8iniJl~ required to 8ccaPgli8h tb0 QpMIfIc doti88<br />

Of b18 iW8diAt8 job. Tbls � rpect Ir wr8 concwrd rith tkr parformaxe<br />

on OpQclfIc UQxrw 8rMs of thQ rpecialty or thr uoiqar equipwnt<br />

of hi8 pQrticulQr ualt. In this PhoBQ he � pp1b8 thQ fundaa8ut8~r and<br />

throry 14axmd through tha Curer Mtrlmt Courrer to hi8 QpacIfIc<br />

unit rlrrim. ‘2hIC pb~8Q I8 8up~~is8d Qnd Qccoxplfrhed priMrIly by<br />

the InditFduQl casu.ndo, rfn(es or oven dam to equadroa hvrlr, .<br />

m8 under Our pt8SQnt cwc’rpt end method of devrbpa8nt me rQ1nt.d<br />

directly to thr C8rQQr D8valopwnt CCRirQOQ. They 8~rtQ priMrIly M<br />

and-of-cctrra �⌧���������� � dmInIrtar@d under coatrollrd conditiona. A<br />

rQ8QQrch progrm Is prQmatly uQdQrwQp to QvQluQte thQ SKTo drtrlop~d<br />

to pQrQ3lrl thQ Cuoer Devolopent Cesrra8.<br />

10<br />

il<br />

.


.<br />

__ .<br />

____ . . .- ---__.._ -<br />

__.___ ___ __----.-<br />

.<br />

.<br />

-.- --- -<br />

l%r Air Force l8 ozl the vuga of opar.tia& ufoptfa$ a raw<br />

tecbique of andysing jobe drfch we are hopeful wit1 pravida more<br />

def?lftive fsfozzaatlon on wbicb to baaa both tr&fniag md mote. UadQr<br />

this @yetam task invcmtorler are prepared gad informatiou gathued frca<br />

fob incwhbmtr coucuuing mount of ttsw, hportanc*, etc., relative to<br />

each opocifie tuk. Through the use of high-apeed ceqmtero aad grumplrrg<br />

trchiqaaa, highly useful Lafornatiaa can be obtainad to aorirt in<br />

more occuratrly daaeribing vuiou8 jobr, to plan training curricula,<br />

aud to urite better tear. -<br />

If sithu of them tuo rrccmt devalopentr-tha clorer rolatiourhlp<br />

of trainiug md testing and tba edro~cor ia job malyria--achaally<br />

lln up to mar upectatiom, �� � �� rO8t optimfctic a8 to the future<br />

of the *trman terting progrrn in tha Air Force.


,<br />

I<br />

.<br />

,*<br />

.<br />

. -.<br />

- - . . ___<br />

.<br />

. . -<br />

.- .-_<br />

.<br />

�<br />

Raaarkn to<br />

Wlitary Teotihg Aooociatfon Conferanco<br />

CAPTAIN J. P. MARTXN<br />

CoexamdFng Officer<br />

US Coaot Oua:d Training Center<br />

AD you know, leot year’8 conference we6 held at the Coast Guard<br />

Training Station at Groton, Connecticut. It was hooted by the Coaot<br />

Guard Inotttute, undrr the c-d of Captain Kurcheohf, who woo chairman<br />

of that conference.<br />

Since then, a number of changea have taken place in our orgnnirutioa.<br />

Captain Kurchukl hu retired; tho Inltituto lo no longer an independent<br />

c-d but ha8 been ccabined with the training rtatfoo to form what io<br />

nou knoun a8 the Coaot Guard Training Center; uumorouo changea in parronnel<br />

havo occured; Llrutanent Camander Hallock, for example, hoe bera<br />

rruoigned � o Ccman ding Officer of a cutter out of FJeu Bedford. In<br />

fact, of thoom uho attended lart year’8 confareaca, Liwtanaut Williir<br />

the only return-.<br />

For the raot of uo, we are attending an Wu Meeting for the firot<br />

timI. But I have had good reportr on laot year’8 mting. By all<br />

account8 it WI a plearant and informotfve confsrente. So I have reawn<br />

to look forward to ‘get.ting bettor acquafnted with you. I (m our0 we<br />

will find many � ��� of c-n interect. We may find that we have<br />

� imilar problems which thir confw-ante mfght hprlp uo oolve.<br />

So on behalf of myoelf, the moaboro of my otaff who ara hare and<br />

officer8 reprorenting other Coaot Guard Unite, we are glad to be here,<br />

to joia in the dlocuooion , and to have a chance to develop ruu inoighto<br />

in tha field of milltaxy teoting.<br />

A year ago 1 wao oboard the Coaot Guard Cutter Northviad, conducting<br />

oceanographic � urveyo off the coaot of Horthero Siberia, and I can � ooure<br />

you that my oula concern8 at the tbaa wore far rroved from the field of<br />

tuting. If � meone had mentioned a “correlatfaml cluoter,” I night<br />

have thought it had o-thing to do with the Aurora Boreallo. And if �<br />

boatewaln’a mate flrrt claoo had came to ma and aoked, “Cap’s, how ccsme<br />

I didn’t do no better than I done on the Chief’r Exm?” it would never<br />

occurred to ma to inform hla that ha ��� � uffering fra a “regreooion<br />

� quatfon” and had wound up un the wrong old. of “an unrrlectad o~c~ooo<br />

ratio.”<br />

Throughout � rorvico career, of course, wa are apooed to a wide<br />

variety of teotlng; techalquao- but uwally in the role of victim. So<br />

it is a nova1 and intriguing experience to otmd “on the inride” at<br />

lart-to mope within � charmed circle whore one hoar0 mention of “*ufltlvariate<br />

mothodo” aad “attitudinal framer of raferenca.”<br />

.<br />

. .<br />

,<br />

12<br />

. ._____ - _-_.- -. -.-. ---.-----. ..----. .--____ _ ..__<br />

.<br />

.<br />

. .<br />

.<br />

.<br />

1 .<br />

_-u--w- .___<br />

.<br />

.<br />

V<br />

.<br />

.<br />

I


.,<br />

� �<br />

. 3<br />

.<br />

i’ i<br />

i I<br />

,<br />

‘<br />

.’<br />

But ot my rate, at tba Cacut Cuerd Ttaiuing Canter I have 6 gaod<br />

crev of mprrtr who havr bean putting out sang thouaanda of erxuoinationa<br />

over the yeara--and I aa sure they have has; accoepliehing their purporo<br />

@it8 Wdl.<br />

What 18 our pwpors ia conducting czaminatfonr In the Coaet Guardt<br />

!<br />

1<br />

/<br />

j<br />

/-<br />

8<br />

.<br />

Well, before m CM an-r that, it lo necearay to explain that wa hew<br />

ravoral dtffarsnt kinds of auminatim progrrrais. And thin calls for a<br />

furthar word on tha reorgrniutioa � t the Training Center.<br />

What hu hawned la tbio: Wo Coast Guard Training Station md<br />

the Coast Guard Institute--which ware two Bepareto ccmun da-ham bean<br />

eooblned-a8 I 8ald srrller-to form tha Coast Curd Training Center.<br />

I<br />

1<br />

I<br />

i<br />

,<br />

I lm urigmd a8 C-ding Of fleer of the Training Cmter. A s 8ucb.<br />

I report directly, to the Ccmtanbt of tba Cout Gumd in Warhingtoa.<br />

‘1 \<br />

.<br />

8<br />

.<br />

Tha Coast Guard Training Center porforma tbrea priaary mirrionr.<br />

Thr Inatituta baa fu?-nirhed tvo of tha primaq Th8 n.taaion diVi8ioa8:<br />

Io8titute Division for Correrpoadcoeo Couroea and the Buairmtioo Divi8ion<br />

for Servica-Wide Examimtion Progrw.<br />

The third prisury misaim divlalcm of the Training Cantar is the<br />

Romfdmt Training Divirlon. This division contiauod to perfore the<br />

dutier of what -;~a formerly called tha Training Station.<br />

Ihe Bcrident Trafniag Dfvfrion camprim & mmber o f technic4<br />

8chool8 which afford trn1nFng to a COntk’tUOU8 flow of atudsnta Who cOQO<br />

in from 811 over thr Coast Guard. At ray one time the atudmt body will<br />

aumbrr about 700 ma. ~hep learn hou to care for aida to nrvlgatiou<br />

equipnat, rcmging frm unlighted river bouyr to rm’s@ lighthourw.<br />

Thay praparo to operota Loran Stationr with 13O%foot tranolnitting towera<br />

rhlch � ra atrung fraa the far reach.8 of the South Pacific to icebound<br />

prmontorier north of the Arctf: Circle. They lcara to baka bread, to<br />

road the dft8 and doha of %x83 Coda, and they acquire & great variety<br />

Of other 8kill8.<br />

The Inrtitute Dlviaion vaa 80 nmwd in order to retain the term<br />

“In8tituta” which haa had an honorable hiatory datfng back to 1929.<br />

Over thr yurr it h8a brctano knovn to hundreda of thouasada of Coart<br />

Cuudrmen asrking advracment with tha aid of corrrapondmco cuuram.<br />

for<br />

man<br />

Th. In8titutO DiVi8lOII pr8plre8, priotl, and 8dlBfni8tcr8 cOur8.8<br />

nrrrly all cO88t Guard 8nlir-d r8ting8: frm aviation � lectronicrto<br />

yaunn.<br />

*<br />

.<br />

,<br />

.<br />

.<br />

‘.<br />

* *,<br />

\ :<br />

/* :<br />

:<br />

t<br />

,<br />

I<br />

I<br />

i -<br />

I<br />

.<br />

I<br />

,


--<br />

.<br />

. .<br />

.-I)-.-. _ _ _ .__- __ _..<br />

_- --<br />

At tha praoent t&ua � hundred and four coutaea ara ia the fiald.<br />

Ap addltional twenty couroeo are being developed. Naarly all ccw~oeo<br />

am undergoing conotant rwioicn. Currently, the enrollment overagea<br />

about 13,000 � tudento.ac any ene time. The average lesson oubo~looion<br />

rata IO .vory cl006 to ona per month par cnrollea.<br />

The preoent Director of the fnotitute Diviolon io C-nd8r Dohlbp.<br />

tilLa Ioyoalf, c- der Dahlby hao recently caspleted a tour of oaa-duty<br />

md he tello me that they didn’t talk much about “correlatlonol clustar8”<br />

on hio ship either. Having some fraa B coosand of I bouy &wider. tha<br />

Cutter Salvia, ho oeosao to think thot “Pcotingrr’o Notion of Cognitive<br />

Dlooonance” hao � uoethlng to do with fog � lgnalo. So I would � ay that<br />

it is juot � o well thot wa both are here.<br />

Both the Reofdant Training Divioion end the Inoeftuta Divlolan uoe<br />

wny quroriono ao laooon aide. Ttmy conduct claooroom and end-of-courao<br />

�⌧����������� � Thora exemo ladlcato tha degree of ouccaoo in completing<br />

a limited couroo of � tudy. They are traditional teaching devices.<br />

But the divioion of the Training Cantor which I would think lo moot<br />

ccmcbmod with tho aganda of thio conference lo the Examination Dlvioion.<br />

Thio divioion io under tha Dfrection’of Couxaandar Turner, yet anOth6r<br />

rljCent arrival. Mloo Turnor’o Diviofon lo primarily engrgsd in the oort<br />

of tooting that wo are here to talk about.<br />

the Examination Divirion conduct8 � rrvica-vide (that IP, Coaot<br />

Guard-wide) oxominationo of milttory poraonnol. Their usage io go>->rnod<br />

to a lorgo Bxtant by th8 manpwor noodo of the Regular Cooot O&ard to<br />

prrfotm ito poaco iaa mloriono and by the anticipated noodo of a grutly<br />

apondod Coaot Guard in tho wont of cobiliratlon. Peacetime miooioar<br />

rmain fairly c’onotant, 80, accept for updating mdtorial to take account<br />

of technological changoo, tha regular -inatIon program tendo to r-in<br />

fairly � toblo.<br />

Bocouoe war plan8 8ra rwioed, uponded, or occaoioaally cut back<br />

to accord rlth internation � graemento. aetloaal policiao, and changes in<br />

military concopto, our tralnlng and examination progrew-for tha<br />

roewver-utuot romoin flaxiblo.<br />

Bocaure of thio apparent dioparity in roquiromento botwoon Regular8<br />

and Raowvoo-and also bocauos of osparata � dm1nLotrative hiotoriro--us<br />

hava, in tho Coaot Guard, developed two diotinct axominatlun programo.<br />

I � hould like to uke a brief camparioon of the tvo.<br />

14<br />

. . . . - .<br />

. .<br />

,.


F<br />

‘.<br />

r2r<br />

. *<br />

. .<br />

�<br />

, ,<br />

! ,.<br />

.’<br />

\ .<br />

.<br />

-_’<br />

.<br />

.<br />

-- -.- -_._<br />

.<br />

,<br />

. -- - _ _..”<br />

..-..-_ ^. _-_..--- _ ” . _ _ . .__ . .._.<br />

Requitemnte<br />

�������� � ra rsquireC to taka rervice-wide axmtinatioue for odveacsm&nt<br />

to the thrao chief ptLy officer gredU --E-7, 8, And g--And for the<br />

firet wmrruat grade,. W-l.<br />

Ra8erven. an the other hand, are required to tAke ~xamLnacioa8 for<br />

advancewtnt to all petty officer grede8. In other word., they ere required<br />

for 3d class, 2GlAec, let clAee md chief--tbAt it pay grader E-4 through<br />

E-7. The Rertrvee do not have E-B end E-9 grade%. Easslinacisnr for<br />

Reserve mdvsncmtr are separate And dietinct fraa the exAminAtiona<br />

� daiaietered to the Regulari. A Raeerve applicAnt for WArrAnt grade,<br />

U-l, hovever, tdcer the sama ex8minAtlon a8 doer � Regular opplicAnt.<br />

hi8 difference in re?uira\lente for exmmlnetione would eppeer to<br />

t&8 into Account the fact that the Rseerve’e path of AdvAncemnt loAdo<br />

throua study couroee, veekly inetructions, and service-vida axunineefon.<br />

By contrut, the RsgulAr’r prospect8 for prmotion have e lerger elemcut<br />

of practical experience, daily performme of duty, And pcraozul evAlu-<br />

, ation by hle caimmding officer.<br />

Both Regular end Reserve are required to cmplete M appropriate<br />

corrrepcmdrnce coureo before bQing conridered for sdvmcentne. The<br />

d;ffarence enter8 tiei tha Resew8 aAplring to ps~ gr8der E-4, E-5, 8nd<br />

E-6 met not only rucceeefully complete an end-of-course oxem, but must<br />

aleo canpete in a service-wide mu&nation. life end-of-coures exam fe<br />

cdainirtered by the Inetitute DLvieion; him rrrvice-vida -inAtion is<br />

� dainietared by the Examination divirion.<br />

Prequancy of Examirution6<br />

For the Regular: Half of the CPO rpccisltfar Are exesized in even<br />

ycure. Thr other helf � re mined in odd yurr. WArrant officer exam8<br />

are &ministered during odd yeer8 only.<br />

lot the Reserve: drrainatione in all enlirted rpecioltier Are given<br />

twice each yur. Inrofar 88 posrfble, there -inAeiOn6 are bared on<br />

Hovy Exmliartion8. Rxminatfone for Advancement to warrant (U-1) Adnfn-<br />

Laterad once each year, we the 8eme a8 for rtgulArs.<br />

Prepsration of Examination6<br />

ExAminAtions for the regulAr program are caapltttly revised each<br />

time they us givm. hey conrirt of new itans furnished by itca-writer<br />

epeciallste brought to the Training Center on teeporAry AdditionAl duty<br />

prior to each group of txeminAtion8. Thair offering8 Are caabfaad vfth<br />

itar eelecteC from M item burk. Cur bank of quertione nw contAia8<br />

approximately 32,OQO ieaar of carefully guarded krowladge.<br />

,<br />

.<br />

,<br />

.<br />

I’


\<br />

i<br />

i(<br />

.f<br />

I<br />

.<br />

As for the Reserve examlnatim program: Xn 1953 Bupers authorised<br />

the Coast Guard to use Navy examinations in our Reserve Progress. These<br />

were administered frua Coast Guard Headquarters until 1962. At that time<br />

administration was transferred to the Coast Guard Institute. The Navy<br />

exama are adapted to Coast Guard use insofar as possibke. Sots8 Navy<br />

exams are not adaptable to Coaat Guard use--notably those for storakeeper<br />

and ye- since our qualifications for these ratings differ from those<br />

for Navy personnel. In addition, we have certain mergency service<br />

ratings such as Coastal yorcarun and Dangerous Carganan which not only<br />

have peculiar names but are peculiar to Coast Guard mobilization requirements.<br />

For these we must prepare our own exam1natlons.<br />

Finally, Scoring and Evaluation Procedures Differ:<br />

Exams taken by regular applicants are scored manually at the Coast<br />

Guard Training Center. Item analysis is also done manually. These exams<br />

have not been adapted for scoring snd analysis by data processing equipment.<br />

Exams aclministered to the Reserves, on the other hand, have been<br />

adapted to the Navy data processing equiprent at the Naval Emaining<br />

Center at Great Lakes. Ue obtain scores, item analysis, and exam analysis<br />

by means of twice yearly visf ts to Great Lakes. There Is, of courue,<br />

greater volume of Reserve exadning than of Regul8r.<br />

This then is the outline and ssope of our esminatlon progrms au<br />

conducted by the Exsmination Division.<br />

On basic purpose in giving exmainations IS to obtain an objectiva<br />

measurement of military and professional tmowledge to use as one of the<br />

factors In assisting in establishing an order of eligibility for promotion.<br />

By canbining the results of comprehensive examinations with other evaluation<br />

factors, we hope to find the persons who are best qualified for<br />

promotion.<br />

Service-vide examinations also give us a clue as to the overull<br />

effectlvenass of all our service training programs. As we standardize<br />

our methods and procedures ve hope to be able to ccunpare our training<br />

achievements with those of other services.<br />

I think ve can do wre to standardite our requlremzntu for both<br />

Regulars and Reoervas. Only by crsatlng a parity betveen Reserve<br />

rrquirements and Regular requirements will ve be able to judge the<br />

relative veakncsses and otrangtho in our tvo kinda of training.<br />

The diffarances aistfng between Regular and Reserve requirements<br />

,for examinations are currently under scrutiny by tha Coast Guard. plans<br />

.<br />

+,<br />

c<br />

‘ .<br />

*.<br />

.<br />

.<br />

.-.-._ - _<br />

.<br />

I


-<br />

;<br />

,<br />

. -_<br />

. .<br />

. .<br />

.’<br />

,<br />

i:<br />

‘.<br />

:<br />

‘. :<br />

. .<br />

. .<br />

.<br />

.<br />

.<br />

.<br />

ate afoot to bring the two examination programs into closer altgnnent.<br />

It is hoped that through the reorganization vhich is still taking shape<br />

at the Training Center, we will be able to streamline ths execution of<br />

our prarsnt callnitnents. And then we vi11 be able to assume an expanded<br />

� �������<br />

The big questions remain--and probably always vill.<br />

Bow vell do we measure what we are trying to mcaaure? What do ve<br />

wind up meaauting after all?<br />

Are present techniques actually reletting for us the kind of men<br />

we vaTin positions of trust and leadership?<br />

Can ve ever hope to accomplish thin goal with “objective” testing7<br />

If not, to vhat ext@nt are v6 “6tNCtUring” our group and our<br />

rociety with an Inadequate tool?<br />

Perhaps if ve could find the kind of person ve are looking for ve<br />

wouldn’t have to bo EO concerned with “pro-pay,” “fringe beneffts,”<br />

and “morale.”<br />

So once again, we of the Coaat Guard are glad to be here.<br />

We shall look forvard to beneficial changes in our “attitudinal<br />

framer of reference”.<br />

We hope that our “regression equation” canes down.<br />

(And I cartaialy hope that<br />

Thank YOU.<br />

.<br />

17<br />

ie a god thing to hope for!)<br />

i


.<br />

i<br />

. . . . .<br />

’ .<br />

.<br />

! --_ _ .<br />

I 1 . --- - . .._. _.___ __ ._ _- ____ _..._--_ - ______.. _ -.___ _-__.._ - _. -.. ..-__---.----.- _<br />

.’ ,<br />

.:.<br />

._<br />

. .<br />

.<br />

,a’<br />

i<br />

i<br />

lkmirk8 to<br />

<strong>Military</strong> <strong>Testing</strong> A88oclation Conference<br />

C- JAMES C. DONAGDEY<br />

Can6undlng Off tcer<br />

US Army Enlinted Evaluation Center<br />

It io with pleasure and a 6en6e of pride, 88 the C-d:Lng Off icar b<br />

of the US Army EnlFated Evaluation Center, that I have thir opportunity<br />

to present to you the asresrrment of the Army’8 syrteaa of maanrring anlirted<br />

proficiency. As you are the experts, I 0111 leave the technical otpOCt8<br />

of job proficiency testing and evaluation in your able hands and X will<br />

u8e the opportunity of thi6 brief prarentatioa to:<br />

Di8cu88 the progrsns of our Enlirted Evaluation Syotae, with p6rti-<br />

CUlar UaphaPir on the nev prOgrsln8 that have tripled the 8COpe of operation8<br />

of the Enlirted Evaluation Center since the lart <strong>Military</strong> feoting<br />

A88ociation Conference;<br />

Pre6ant rigntficant changer -da in operating procedures hara at’<br />

the Center and a new method of documenting and reporting tndtividual<br />

proficiency; and<br />

To point out the team effort that ir required for the davelopmant<br />

of avalu6tfon material8 to rupport the objective8 of per8oWel wemant<br />

arli the importance of 6ttaining there ObjeCtiVe8.<br />

The queltty coldier ha6 been sought by our Army 8ince its earliert<br />

inception; although mo6t American pioneer8 Vera proficient fn the u6e of<br />

the musket, it VP8 not 866~ to find proficient soldie- even ia there<br />

&y8 when a limited knowledge of technlcal material was required. The<br />

quart for quality m6npwar has continued down through the yearo. Wa all<br />

certainly would agree that the way to proficiency h&r to be the quality<br />

men -- the aright t man t hfor e the rright i gjobh t t i m e .<br />

A6 Commander of the Enlirted Evaluation Center, my raiseion i8 to<br />

develop and monitor technique8 for evaluating the occupational proficiency<br />

of enlisted perronnel in the Army -- to maiutain a field<br />

administrative syetan for operation of the program -- and to render<br />

ccntralited 8COring and reporting of the rerults of the terto.<br />

Fran obrcurity to the vorld’6 second lergent testing center in 8iX<br />

rhort year8 16 the outstanding achievement of the US Army Enlisted<br />

Evaluation Center. Since March 1958, the Center ha8 developed from a<br />

aina+aaa operation, located in a “borrowed” ClaS6 ro~ll, to more than<br />

200 military and civilian parronnel (not Including the field te6t control<br />

officer8 throughout the vorld).<br />

_ _..-_ . - . . -<br />

F I’ .<br />

.<br />

1 ,<br />

: * .<br />

1<br />

�<br />

18<br />

. .<br />

c<br />

:<br />

:<br />

-._-__-_- -.<br />

.<br />

.<br />

.<br />

���� � �<br />

�� ��<br />

.


‘.<br />

<br />

.<br />

The initial objective of the Enlisted Evaluation Center in 1358 vas<br />

to evalua: s enlisted personnil ln their primary MOS for award- of proficiency<br />

; s;.<br />

The succtsoful integration of this program into the Army’s systaz of<br />

personnel managcmcnt prompted adoption of additional programs that could<br />

be administered and controlled through the enlisted evaluation systezz.<br />

e Cur system now supporte--in addition to the proficiency pay program-the<br />

following functions of enlisted managcmcnt:<br />

.<br />

h<br />

Primary MIS Qua~ificntion.<br />

Qualification for Promotion. Although this is used on a pcnnissivc baais<br />

within the Army, the system dots furnioh the coumundcr an indication of<br />

the soldier who is best qualified to fill a position of responsibility at<br />

a higher gradc.<br />

Secondary and Additional MS Qualification. This function provides commanders<br />

and Dtpartmtnt of the Army the information required to tfftct<br />

broader utilization of the soldier by conridcring areas other than primary<br />

job in which he is qualified. It allows use of tht vholt man instead of<br />

part of him.<br />

Reserve Cocnpontnt Evaluation. This ncv program is aimed towards the tvaluation<br />

of our citizen soldier, the Army Reserve and the Army National Guard,<br />

on the same standard as the Activa Army and vi11 assure the use of the<br />

ssme means of measuring results toyards tha aamt goal--a better soldier<br />

throughout the Army structure. Thie is one more important otcp towards<br />

the attainment of a one Army concept.<br />

The Enlisted Evaluation System has answered the Army’8 need for an<br />

objective system of individual evaluation of enlisted personnel co support<br />

there progrsms.<br />

Prom a standing start of about 17,000 evaluations in 64 Army jobs in<br />

1959, the Ctnttr vi11 evaluate almost one mfll1on personnel during fiocal<br />

year 1965. Thfs will require publication of approzimataly 300 ttst aids<br />

and the development of over 1,000 evaluation tests.<br />

Our tests provide the commanding officer with the information with<br />

which to asssss the capabilities of the Individual roldicr assigntd to<br />

his unit.<br />

For example, a tank commander knows that the tanks aszigned to his<br />

unit have a ctrtain maximum speed, destruction capability, and crufsing<br />

rcngt . Likcwiet, an artillery battery commander knows that the guns<br />

with which he is equipped have prescribed muzzle velocities, ranges, end<br />

rate8 of fire. Not only must he know these specifications of his material<br />

but he must know the particulars of tht jobs employed in his unit to<br />

19<br />

,<br />

,<br />

.’


I’<br />

I’<br />

.<br />

-.<br />

_--. .- _<br />

. _ . ..__. ?..-. ..___<br />

.<br />

8ccowplieh unit lgiaaiona and the capabilitie8 of his m to perform<br />

these joba.<br />

The eoliated avaluhtion ayetesn provides the cocmendoro with 8 more<br />

factual basis for umting personnel mamgement decisions.<br />

The impact of the ~rmy’a tmlirted svalu,rtion eyotdlrn on management<br />

can be exprsoaed in three aroaa: Pirat, it afford8 enlletocl peruonnal<br />

motivat1oa for greater job m88tery, maintenance of currency in job proficiency,<br />

and la a morale factor oiace peroonnel d8ciaiona c:an bs baaed<br />

on objectively appraised writ. Second, it furniehoa tho unit comundar<br />

with 8n objoctivo muaure for 8ppraioing individual cunpetonco, for<br />

daterP1ning thoao poraonnel beat qualified for prcmotion, � �� for uoo<br />

la araigment 8nd utlliution of his peraonnol 8od highlighta tequitomento<br />

for individual 8nd unit training. Third, Depettment of the Army<br />

ia provided vith effective me8auroa fot implamonting personnel m8nagowont<br />

programs and 8pplying uniform � tand8rda in peraonnel rmnogammnt.<br />

Direct’aupport of theao paraqmel wnagownt functions la provided<br />

by the Ealiatod Evaluation Center.<br />

Since tho last conference, the Centet has achiovod 8 aign1ficent<br />

� cv8ncaneot in the method of documanting and reporting individual proficiency.<br />

Inotollation of tho Center’s cozuputet � yateta provided the<br />

capability for instituting a total redesigned evaluation report. Pl888Q<br />

tefet to the copy of the boos Evaluation Data <strong>Report</strong> which ha8 baon<br />

included in your brochure. Prir.: to canputar epplication, individual<br />

proficfency was reported as a single numeric rvaS.uatian 8core ma indicated<br />

in the upper right hand coroer of your sample report. Nw indivfdual<br />

reports of HO-S evalu8tion not only ahw tho axamineo’a attained evalu-<br />

8tion score, but 8180 his atrongtha and wcakneoaea in the functional<br />

8raa& of his occupational apocislty as shown in the lower portion of<br />

your amp10 report. Your aauple ahwa tho seven subject 8re88 on tho<br />

Evalu8tion Data <strong>Report</strong> for an infez:: y senior � erge8ot. XOS 115.9. A<br />

det8iled derctiption of those areas is given in the Evaluation Teat Aid<br />

for thia HOS. Sgt Baacomb lo Very Law, Lov, or Typical in a11 are88 of<br />

his 190s. lie and his unit would benefit from his study in 811 theao<br />

� ���� �<br />

Copier of this report are forvmded to tho examinee’a unit of<br />

8aa1gnmcnt and the individual concerned. Special added di8tt1bution<br />

la rude to DA, where results are used to determine various personnel<br />

� ctionr �<br />

With this Evaluation Data <strong>Report</strong>, corman dare can identify nacearary<br />

training requirements, on-the-job, and thoao � ubjoct-xaatter are88 in<br />

which form81 school training la aaaent1al.<br />

-_- .-.... ___-__-.<br />

- . --I-.---.<br />

__ _. -. .- -_ __- -<br />

.<br />

20<br />

. . -.<br />

.<br />

.<br />

,<br />

I<br />

--. c ---- -- ---_l-___<br />

.<br />

I<br />

,<br />

.<br />

.<br />

.<br />

1__1___<br />

.<br />

.


\I’<br />

. .<br />

i ’<br />

’ _<br />

. .<br />

I<br />

1<br />

Z-N -A*<br />

.<br />

*<br />

�<br />

.<br />

__. -- - ._ _-.. - ._ . . -- -. . -_-- -_._ --. .--_ --~_. ..^<br />

. .<br />

. .<br />

.<br />

.’<br />

�<br />

-...-. .._---.. -._- --.--. ----.- .-.-- __ -- .- _ ..__ _ - --..<br />

.<br />

-.<br />

At this point, f might add that la past MTh Conferences dlscusalona<br />

were hold concerning the cxpericnccs of the othl:r servlcce with proffllag<br />

evaluation scores. As a result of this exchange of lnfornation, the<br />

Ealfrted Evaluation Center va8 aided in ltr efforts to produce A Profile<br />

Data <strong>Report</strong> which accaepllohe8 those purpocea unique to,, the Army.<br />

SWry reports orb dlrtrlbuttd to -jot ccr=eandcrr in the field.<br />

Those report8 reflect tha’WS, tbe rpread of evaluation scores, And the<br />

total nunbcr achieving each @core by grade and unit. With this lnforrrmtion,<br />

cameender8 can ApprAt80 unit efftctlveness md emphu~lre training<br />

requlr-nts.<br />

The objective of Army Perronnel Kanageaeant IO rimply to obtain the<br />

uulm~m cf flclent uee of raenpwar. To oupport the att8lnment of thle<br />

objective through thr aunagesent programs I hava cited the evaluation<br />

process requires the coordinated effort of the Center, vhich is the<br />

oparatlog Agency of the ry8tea A group of 10 Auparvlrory and 31 question<br />

writing ~genclas who ftrrnlsh lndivld~al question8 And problms which<br />

maka up the various tests, and 81 monitoring c-ds who ~upcrvlsa and<br />

8UppOrt the actual operation of the tcitlng program at the many far flung<br />

locations throughout the world jlere soldiers are located.<br />

The drvelc+ent of evaluation material8 fs a team effort. Toot<br />

speclAlists here At the Center are in direct c~nlcatlon with thelr<br />

carnterparts at the question writing � �������� Thir te4un affort Insures<br />

that profesrltmal rtmdards of test develomt ara ir,;orporeted In the<br />

8yStam. Through the use of Appltcstlar of measurement prioclplos by<br />

professioual pcrconnel the Army’8 evaluation lnstr..srsntc are in liar<br />

with those rtandatds employed in professional prychologlca:l maa8ur@aeent.<br />

Thlr later-action batween the Center And the ochoolr Is neccrsary to<br />

8U8tAln the highest profesrlonar. standards in our devclopnmt of telt<br />

Aid8 cad tests.<br />

Distinct advantages ACCNO frau this tt)za effort. For ermple:<br />

It insureo inclusion of accurate and current: subject-mrtter in the taste<br />

by the question vrltlng agcnclcq it provides for � ppllcatlon of momsuremant<br />

prlnclplc8 by the Center’s proferrlonal parronnel, and thib in turn<br />

Insures that the test masuras the techaicel knowledge required of tha<br />

roldfer to rdoquatoly perform on hi8 job.<br />

The military mvlronment to&y is still one of the moat caaplu<br />

social organizations lu the country. The problgn of finding the right<br />

man for the right Job 18 A continuously expanding task.<br />

.<br />

21<br />

�<br />

. :<br />

.<br />

~-- --<br />

i i<br />

!<br />

i<br />

I<br />

t!


.<br />

.<br />

-- .<br />

� �<br />

.* ,_ _._._.-_ _ : ,.. : _<br />

. .”<br />

.-<br />

“G, i .<br />

.<br />

.<br />

.<br />

L<br />

Yhir ck8rification and aroigment technique involve8 all of the<br />

farm of mmagesumt fmiliar to em& of you, that 18 tha identification<br />

of aptitudar, laarued � killa and knowledge, qualification for odvancemeat<br />

and relectlcan of peraonnal for opeclal � �����������<br />

Our preuent � y8te6 ha8 conridereble merit since it is producing<br />

the right man for the right job.<br />

I believe the 8electicm nystcm, to find the right man for the<br />

right job, should be dafigltlve In purpo80, yet r-la adainl8tratlvsly<br />

r-10, and above all, arrlot the cmdar in the oelection proceru.<br />

You ladier and gentlawn realira the awunt of work and proferriorral<br />

application required to dovelop the tcating lnrtrumaatr to achieve thir<br />

objective.<br />

In my opinion, we have an excellent reearurercant devica to identify<br />

current lamvledga on what the men “ten do.” Rowever, soort ‘mult be<br />

accanpllshed m the ‘VI11 do” � ppllcatlon of the individual onca ha ha6<br />

been placed on the actual job. We muat not lore the confldance of<br />

camander8, vtroor we tell that the i?dlvidual PO job qualified, aad then<br />

have the mn fall mirorably when he is arrignrd to the job.<br />

I have examined the topico on the program for dtccurricm which I<br />

find nort thought provoking. I look forward to the opportunity to rtudy<br />

your dircurrlon group report8 end benefit from the trmrendous collective,<br />

technical end p~ofesriozul knowledge represented by your presence. The<br />

Eallrted Eveluation Center vi11 continue to rupport the Asrociatlon’~<br />

effort8 toward future lmprov-tu.<br />

Let me say once again that the US Army, and partlcu1arl.y the<br />

Enlirted Eveluatioa Center, � ppreclate8 the opportunity to hart thio<br />

Sixth Annual &CIA Ccmferenco at Fort Beujmia Harrison.<br />

Ihi concluder the formal oerrion tod.ey. I vi11 be followed by<br />

Mr. Price, our Conference Chairmm, who ha8 8ome announcement8 to make.<br />

22<br />

---- -- ---


. ,<br />

. .<br />

Perspective<br />

DR. HAROLD A. ED!XXION<br />

President, Performance Research, Inc.<br />

Washington, D. C.<br />

First of all, let me express my thanks to the speakers this afternoon.<br />

They have done a beautiful job of saying,in a different way, the<br />

same things that I want to say again tonight.<br />

I am listed here as the keynote speaket, and I have ‘been watching<br />

keynoters of the political parties this 8Umner to see just what they<br />

do, 80 I vi 11 know what to do tonight. First of all, th’ey seem to look<br />

at the glorious and constructive record of the past--what we have done<br />

that is notable, forward looking, and looks good on the ,record. Tha<br />

second step is to paint with dark colors the errors of the opposition<br />

and to offer many castigations for these errors. And thse third phase<br />

ie to point out the great and bounteous future which we offer. I<br />

propose to follow this as an outline this evening. It gfves us a good ,<br />

framework within which to take a look at our own problems. I have<br />

chosen the term “Perspective” aa a title. I thought thins was sufficient<br />

to offer maybe some direction, maybe some closures.<br />

To keep this keynote speech and its implications clear, we need to<br />

be sure that we are using some of the same points of view and some of<br />

the same definitions of terms. Let me define the term “teat” aa ueed in<br />

our discussion this evening. We are a group primarily ci=rned with<br />

the construction of achievement tests--testa which ahov job knovledge,<br />

understanding and skills. A teet, as we are using the term here, is a<br />

sample of behavior, dravn under such condition8 that cne may judge some<br />

set of skills abilities, aptitudes, attitudes, or achievements on the<br />

basic of this sample of behavior. <strong>Testing</strong> then becomes a problem of<br />

constructing tests which make it possible to draw an appropriate sample<br />

of behavior. We are concerned with the adequacy of this sampling, and<br />

with its reliability. Are the sampled behaviors percfnent to our purposes?<br />

Then there are questions of uniformity of sampling, from time to time,<br />

place to place, and group to group.<br />

his concept of test ae a sample of behavior ia a very useful concept.<br />

It reduces the %.agic” aspects of tests. Nothing seems mysteriotre about<br />

the idea of sample. Whether you have had a course in tee:ting ata university<br />

or whether you have sat at the feet of one of your technicians and had long<br />

discourses on the nature of tests, you know that when you draw a sample,<br />

there are errors of sampling. You know that you have to be careful what<br />

you are sampling. You know that you have to draw a large enough sample of<br />

behavior, so that it vi.11 represent what you are tryfng LO sample. I am<br />

including within this concept of a test all of our procedures for drawing<br />

samples of behavior. I include here the sort of things we ordinarily<br />

23<br />

\ b.,<br />

.<br />

.


. .<br />

.<br />

. _ . - - ._ ._ -.’ _ _ - _ _._. _..._ _---_. .__ .-. _ -__-.<br />

include as teste. I also include the performance of work sample type of<br />

teat. I include the subjective teat; and we too often look down our<br />

nodes at such meaaures at times, primarily because they have greater<br />

sampling errors under certain conditions than do other senples of behavior.<br />

I include an interview as a sample of beh,avior. I include ’<br />

anecdotal reporto as samples of behavior. <strong>Report</strong>s by superiors on the<br />

performance of subordinates is a report of their sampling of the<br />

subordinatea’e perforaancc. Even projective tests are rsmples of behavior,<br />

but it takes someone well-trained to tell you vhat he is ssnpling, how<br />

this sample is comparable to any standards, and what to do vith it. When<br />

we look at testing an sampling, this point of view also subordinates the<br />

techniques of testing to the real role oftarta,thnt of measurement.<br />

Techniques may take their proper place as voym and means of achieving the<br />

purposes of our tests.<br />

. --- /<br />

c<br />

. i<br />

And nov to look back briefly on our achievements. Our glorious past,<br />

from the point of viev of this group and this meeting, is one to which I<br />

can and do point with pride. I know of no group that has done aa competent<br />

a job in achievement testing. This statement is supported by the behavioral<br />

facts oi the membership of this group. The membership of this group has<br />

advanced the otatc of the art on all fronts -- job analysts, task analysls,<br />

job content, frequency, and so on -- the boscs for teat ccnstruction. They<br />

have improved our techniques of item construction, relating content to’re.-<br />

.<br />

quirecuent , and adjusting difficulty to appropriate levels. They have advanced<br />

the techniques of item analysis; they have also promoted the ncc.:rsity<br />

for item analysis. They have developed techniques for producing alternate<br />

,<br />

forms of teata. Por theae and for their other achievements, I salute the<br />

members of this group, and I take pride in being here and feeling that I<br />

I am really one of you.<br />

’ C’<br />

Then we come to the oppositfon, but the oppoeilton La hard to find, to<br />

identify. Who is our opposition? Are they the ones who refuse to use our<br />

products and try to make tests do thlngs that they were never intended to<br />

do? Arc they the ones that insist on using less adequate mean% of performance<br />

measurement than are readily available to them? Perhape ve are<br />

ourselves our own greatest opporiticn. That is a dangerous statement to<br />

make after pointing with such pride to the achievements of this group. But<br />

among us, I suspect that each of us, including your speaker, has been<br />

guilty of being his own greatest opposftion. When we see testing in terms<br />

of only one set of techniques, we are miaeing the boat. At least we are<br />

putting a good hole in its bottom so it can sink with us. When we fail to<br />

base our testing on adequate analysis of per.Cormance requirements, ve are<br />

damsging our program. When we fail to recognfte and state the specific<br />

purposes of the test in operational terms, we are running the risk of cutting<br />

our clvn professional throats.<br />

We want to know what it is we are trying to mecrure. You do one kind<br />

of job when you are trying to measure the status of knouledg~e. You do<br />

another job when you are trying to bring in skill and rcaronlng, and you<br />

_-<br />

.<br />

.<br />

24<br />

___L_._-.- .--.-_ -- _- _- _ _ - ._<br />

. .<br />

.<br />

. .<br />

.<br />

‘,<br />

.<br />

.*<br />

.<br />

i<br />

.


. .<br />

.<br />

. .<br />

.<br />

need another sort of test item when you are trying to measure memory for<br />

obscure facts. There ar,e times wher each of these purposes is pertinent,<br />

but do not USC them blindly. Use them knowingly. It. brings to mind one<br />

definition of a gentlemen --, one who never unwittingly offends another.<br />

We need to know how the test vi11 be used; we need to know what kinds of<br />

persona will take the test: ve need .to know where and how the test takers<br />

learn the material covered by the test; ve need to know these as part<br />

of the mission of the test we are trying to build, and until we know that<br />

ve run a serious risk of being our own best enemy.<br />

Another current form of opposition, and perhaps not as real as those<br />

just pointed out is represented by those who seem to enjoy swinging at<br />

straw men. They single out some of our weaker, less adequate test items=those<br />

that may be viewed with alarm, with scorn, derision, double-meanlng-and<br />

hold up to ridicule ail testing because they found a flaw in our work.<br />

Some openly question the use of tests at all. I have n.>lt found out what<br />

they propose to substitute for testing, but I suppose there must be something<br />

from the “good old days” which they find adequate. Now when you<br />

look at the kinds of things published by our critics - a castfgation of<br />

tests built on some misuse of test items from such tests as the Bemreuter<br />

or the Bell Inventory, an attack on testing in Fortune Magazine some ten<br />

years ago, and, more recently, a very interesting book b:y Dr. Benesch<br />

Hoffmann who seems to believe that multiple-choice questions are the<br />

bane of our social order -- should we ignore their strident and exaggerated<br />

criticisms? Within each such criticism is some useful information. This<br />

is why I cannot label them as our most dangerous opposition. I personally<br />

value Dr. Hoffmann as my most competent test item critic, He accuses me<br />

of the worst kind of ekulduggery and of a great desire to take unfair<br />

advantage in wrfting test questions. So when WC read these articles and<br />

hear these sp,eches that take us to task for what we may consider a minor<br />

sin, or some error that we have outgrown --- take a second look. We may<br />

learn something, perhaps ,comething useful.<br />

One cannot fight such critics effectively because o.f the many facts<br />

and assumptions which must be righted. But, from them we can see places<br />

where we might improve: how we might build more acceptable programs, how<br />

we might better market our products, and SO on.<br />

The third phase of this talk tonight is our look ahead to see if there<br />

is anything in the future other thsn the manufacture of more tests just<br />

like the latest models we have been nr.king. Recently I visited an automotive<br />

museum. One thing struck me forcibly; there were cars prior to 1912 that<br />

looked like overgrown buggies , pcvered with washing machine motors. But<br />

all of those ca:s built since 1312 were not so old-fashioned. About all<br />

we have done with automobiles since 1912 is improve them,, We have put on<br />

bigger motors and better tires; FJt we have made little real change in them-a<br />

here in a period of 52 years no ,.&al basic change; improvements, yes;<br />

change, yes; but no real breakthr *ugh-<br />

25.


,’<br />

” ._<br />

, i<br />

.<br />

.-<br />

.<br />

.<br />

L . _ _...__ - .- . -. __ _____--,<br />

I think those of us ia the testing business are in that same boat -- _<br />

rod to ray. The development of the format for objective tests, which we<br />

all use and think of quite highly was a product’of A. team like this. We<br />

hAve improved them, shined them up, And dressed them up in different forma;<br />

but We reslly haven’t mcldc A basic breakthrough in ~11 that time.<br />

The AmY hIphA teat used in World War I was A milestone in test developmerit.<br />

That test wan in objective form. It WAN certainly a useful teet,and<br />

it hsd in it many of the kinds of content we currently use. In fact, for<br />

AOCIIC purposer, it is still usable. In fact, it is set11 on the rmrket. But<br />

we build testr for many other purposes. We have used improvements of the<br />

techniquea, but we have yet to realize A real breakthrough. We write items<br />

with one, two, three, four, five,and six AltematiVcU; one of them is right:<br />

the re8t of them art wrong. They did not have at that time the refinements<br />

we have developed in item analysis, using biserial coefficients, or whatever<br />

other statistics you May use at your technical altar. We did not have factor<br />

AMly8i 8, AE A matter of hiStOriCA fact, psychometrics did not reelly<br />

get going until the days of Cm1 SpearmAn, followed later by Thurstone.<br />

The idea of item homogeneity is still A more modem concept. But I don’t<br />

consider there as breakthroughs, They Are improvements in juet the same<br />

way that bAllOn Circa were AU improvement over the old high-pressure hard<br />

tire8, or in the OMIC way some of our modern carburetors are improvement<br />

over the simple carburetors of the pest.<br />

. Despite all this, however, I Am optimistic about the real possibility<br />

thdt we are on the eve of some major breAkthroughr in testing. What they<br />

are, I do not know. I am equipped with neither that kind of vision nor<br />

wisdom. By A breAkthrough, I mean a change or new idea that alters the<br />

whole perspective of test c.onrtruction as well as test usage. Are there<br />

AituAtions and times when another technique would do better; for exttmple,<br />

better As the first step in determining what factors in A job are really<br />

crusial, or really pertinent? Instead of trying to cover the entire job,<br />

maybe there are uome element8 in the job that ere really crucial. There<br />

. were some early flags of that kind back in the 30’s in the development of<br />

job questions by the U.S. Employment Service. It was Asked that these<br />

test quettions discriminate batwern the master or journeyman, and helper.<br />

� Their particular purpose in the employment office uas to identify the men<br />

who would like to upgrade themselves, but who really did not have the kncr:-<br />

/<br />

�<br />

Adge of the mater craftsmen they had worked around. But many had worked<br />

enough so that they had learned something of the job. RAybe there are<br />

Aome clues in using the concept “How do you know?” ‘What did he do?” kind<br />

of questioning. Only once in my career have I velidatled a set of selection<br />

tests All of which measured up to the criterion of performance. Rather<br />

than using ratings of competencgwe tried a kind of job analysis attempting<br />

to set what aupervisorr saw when they rated A man’s performance es “good.”<br />

We asked “Who ie your best man? Whet does he do or fail to do by which you<br />

know that he is your top worker?” We would continue t4e conversation about<br />

one man for 30 to 60 minutes finding out whet he did, boo he worked, how<br />

�<br />

26<br />

. . . -


.___ ---.-.-a ---._- ~.-~“..CI.~.--..w.‘r -.--.-.-^_. -. ._ .<br />

.’<br />

!<br />

)<br />

he handled his contacts, how he cooperated vith the home office, everything<br />

we could about him. Then WC would ask, “gow about one of your<br />

poor ones? Who is.the next one you would drop if you had to cut iour<br />

work force? Haw do you know he is that bad?” We found enormous overlap<br />

in observable performance of the better as compared with the poorer men,<br />

I<br />

a<br />

’<br />

and how he worked, what he did and vhet he failed to do. but there wds<br />

a small area of difference. It did not tell us too much about job duties,<br />

but it certainly told us hw the company valued job-parform=ca.. And<br />

this vas the information that gave us the clues 8S to vhat tests to<br />

select. Maybe thin point of view contains 8n idea we, should investigate.<br />

Instead of looking at all the duties, functions, and rcsponsibflitiea of<br />

the men, try to leern in what wrys we can spot a good one when we see one.<br />

Let us not forget that there is a new set of tools at our disposal;<br />

they are known as computers. Computers and their accessoF equipment<br />

certainly ought to open nev d-ors to us. They have already opened many<br />

doors. The biggest and most important is thet they permit us to do jobs<br />

that were too big to do before. They handle much more data. Before<br />

computers, how big a job of item analysis did you tackle? One hundred<br />

high cases; one hundred low? That would be close to the limit of what<br />

you would try to do. With comp::ters ve don’t think anything of using<br />

revere1 thousand cases in our item analysis. We are tackling problem?<br />

that are bigger erd bigger. Some of you may remember the day, it is before<br />

my time I mfl;ht add, when six correlation coefficients would enable<br />

one to e8rn a I% ;i. degree. As time vent on more correlations and more<br />

complex correlat;.)n functions vere required. We are collecting and using<br />

bigger and bigger quantities’of data in support of our efforts. These<br />

are improvements, but not breekthrougha, because we have been using the<br />

s8me old techniques, but with increased effectiveness, Computers do uuggest<br />

that ve can do some things in ways never done before. For example: One<br />

group of people for vhcxu I heve constructed tests .-or some yeers are so<br />

bright I 8m not able to write items tough enough for them. These are the<br />

science talent perticipants for the Westinghouee Scholare;.ips; these boys<br />

and girls are really bright. Nevertheless, I’m still looking for talent<br />

even in that group. I want a test vhich discriminates particularly among<br />

those in the upper half of the contestants. I want a test which will be<br />

reflected by a frequency distribution that has a positive skew. Some test<br />

questions which come within the proper difficulty level may depend upon<br />

remote and unimportant bits of information. Is it, for example, really<br />

important to knov precisely hw many light years Arcturus is from the<br />

Earth? Or even to know the answer within a thousand light years? This<br />

fs the kfnd of question that is answered correctly by few people, but etfll<br />

did not seem to be just what the situation demanded. Then ve tried multiple<br />

choice questions; but instead of having just one right and 4 wrong answers,<br />

we thought we would have one, two, three, four, or even five right ansvers.<br />

A score of 1 is earned if all the alternatives are marked correctly, and<br />

if there are ;qy errors the earned score is zero, This produced test<br />

questions of appropriate difficulty and gave the right kind of frequency<br />

.<br />

.<br />

27 i<br />

-._ ._<br />

,<br />

.<br />

, 3 t,<br />

^><br />

.<br />

. ,<br />

I<br />

i<br />

1<br />

I’<br />

I<br />

/


, .<br />

distribution. Basically, I have used the simple concept of conpoutd<br />

,<br />

probability,<br />

.<br />

Psychologists and test makers seem to have assumed that all that<br />

one needs to measure, or perhaps can measure, are behavior samples drawn<br />

from individuals. have you ever discussed the validity of a test<br />

question with your field people and had the answer, ‘Well that depends.”<br />

That depends on circumstances of the job, of the man’s boss, of the<br />

working climate available, of the standards to which he is held.<br />

*<br />

Another area that we should explore is this matter of fncludtng in<br />

our factors of performance some which are outside the skin of the individual,<br />

but nonthelaar lnflu5aco the quality of hia pstformance and of hlr knowledge.<br />

These pieces of information should appear on the right-hand side<br />

of the equation; not on the criterion side, We have not explored this.<br />

Since the daya of Carl Pearson we have bowed down in front of linear<br />

measurement. tven out factor analysis is just a complex linear system. We<br />

have not broken with that tradition; we have not developed the tools to do<br />

it. Did you ever try to compute a correlation matrix using nOnlimdr<br />

regression lines? The first thing one does is rectify them to get back to<br />

linearity. You are well aware that there is ao much in human performance’<br />

a.rd human value that is nonlinear. Even starting from mince pies, there<br />

15 such a thing as too little and there is such a thing aa too much. In<br />

between them there f5 some kind of optimal amount. With most of our personality<br />

measures we find the same sort of thing, Perhaps we call them bi-polar,<br />

but we ought to look at these a, porentially nonlinear concepts.<br />

Then we talk about patterns of performance. Wfthfn a set of questions,<br />

suppose there are three for which the answer “yes” are "out." We might call<br />

these “lethals,” Conditional answers also might be explored. If you mark<br />

answer number 3 on question 1, then whatever you answer to questions 13 and<br />

26 is wrong because your prerequisite knowledge was inadequate. There are<br />

a number of these pattern concepts that might be useful and the method tested.<br />

Out computers open the door to such problems because this kind of exploration<br />

call5 for considerable amounts of data. And I might add, your organizations<br />

have sufficient data. You have the opportunity of exploring rationales for<br />

weighting and patterning of answers.<br />

Some of the possible questions are, How good is it for a man in a<br />

given HOS to be a member of the 1 percent who give the correct answer to<br />

quertion 471 Is thic worth more than 1 point relative to the other<br />

questfonr? One might explore the weighting of item5 inversely proportional<br />

to frequency of right anBver5, so that those who have the mote unique<br />

knowledge become more vieiSle.<br />

28<br />

. .<br />

.<br />

* .<br />

-<br />

.<br />

*


. .<br />

-<br />

. -. .<br />

. ‘.<br />

-<br />

-m.__ ’<br />

We need careful definition of what we w4nt aach texst to do. we<br />

IWad ~108tar 6tbt6mant6 of what we want it to ~1~1oura and the condition8<br />

under aiich it owe do ito job. For a given aiammxwnt task w4 may nsed<br />

4 toot which yield8 89% perfect @core4 , or one that nsbodp parscoo ualeoo<br />

ha nsakeo a perfect ocara,aud there ara time0 when thi4 is 4 reoooneble<br />

toquirement. What kiod of diocriaination doe6 your test need to aukt?<br />

Doe4 e4ch queotion naud to opcrata at tha one parcent level? Are you<br />

aokfng it tcs diocrFrninate in ouaa othar fashion? 50 you want the tMt<br />

to be uoed a8 a motivation:1 device -- to moko your teat takoro uu;iouo<br />

to get out 4nd do more, better, 4nd Cantor? Do you want 4 toot that is<br />

basically a teaching dcvica? That wa4 the initial motivr~tion of S. I..<br />

Preooey’o teaching machine. In uoing hio mochineo,ovory time o otudant<br />

puohod 4n answer button to uIewer the tart 4nd got the right 4n4wor, the<br />

-chine went on to tha noxt quartion. If he got 4 wrong rnswar he kept<br />

on prrooing anover buttono until he got tbo right enover,, How vc erll<br />

thio reinforcement; imediaco feed bock.<br />

Thare aro ocane othar quaotiono to which pa Q not h&ve uuwcr~. ilou<br />

volid ohould 4 toot be? We have ha4rd about maximum validity, about lack<br />

of validity; but molly wh4t io the optim41 v4lidity YU rlhould look for?<br />

Anothsr quartion io *How rapidily and under whet cmditicmn do tsot ocoreo<br />

loose thair validity?” I have conoidorad the po4slbility of submitting a<br />

rootirch propooal for the development of tcot record inb: -0 ink4 for cuking<br />

teot recordo -- 80 that the teat � ���� for each teat wou1.d be faded out Juot<br />

at the tlme tha tart 4cora had lort it4 validity. At me time in 4 cwanooling<br />

center, wo act up a hard and foot yle, “Any percouality, intorest, or<br />

oimil4r tert has lort it4 validity 4ftcr 3 months.” After that,o5t4in 4<br />

freoh ocore - 4 04mpla of whrt that individual lo like now. And it might<br />

be noted that people oeeking counoelfag are parh:?o leoot stable in ouch<br />

charoctariotico. For b4ric rptituda teat8 ouch a# tha Ohio Stats Poychological<br />

Tort, we aokod for 4 frarh tart or rumple after one year. We nacd<br />

more knowledge 4bout thfr queetion not only in toner of tho teat, but 4140<br />

the intervening aventr. In measuring how wall 4n electronic6 technician<br />

parform on tho Job you msy get 4 meaouramnt at the time he completer<br />

bio training. Suppore he had �� � it and wait 60 dayo for �� � rafgnmont;<br />

hw volid lo hfo ta4t bcore after 60 d4yr with no practfce in the electronico<br />

ohopt Hw valid would hi8 t44t ocoro bat<br />

In being here tonight, I have triad to point to 4om3 of the future<br />

and to � tir up your thinking. I hove tried to illurtrata the l’taot’@ 40 4<br />

sample of behavior. I w4nt to cheer for axtending 4nd trying out new concapto<br />

of 4n8wer pattarnr in tart fonnulatfon. I want to iace ue devalop the<br />

ure of computer6 to do things in test rcoring and teat analyrfa th4t wo<br />

hove only tolked or drarsled 4bout and to do � ome thing4 that wa h4ve not<br />

avan rtartad to inugine. Keep your w4ys and maano, your rechniques, in<br />

proper peropectiva; and keep in front of you a clsrr idea of the purport4<br />

of the test itself by defining it4 uoeo, 48 well as ftr limft4tion4, 4nd;<br />

through these, extending Its ueefulneoo a8 a me4euring device or w4y of<br />

drawing more 4dequ4te, 4nd more effective samples of behavior.<br />

I .<br />

I<br />

29<br />

_ . _ __ .._^_ - .._-_ . -. -<br />

-<br />

._. ,. .<br />

.<br />

:<br />

I<br />

.<br />

.<br />

*.- ._ ,(<br />

\~<br />

._,. _ _ .____. ._-. .- ___--__<br />

*<br />

--.-<br />

,I(<br />

- --<br />

,<br />

-_-


-.<br />

-. _-..<br />

.- .<br />

- ! e<br />

I<br />

_-.. .<br />

1<br />

.<br />

.-<br />

.<br />

2<br />

-.e.<br />

T-&<br />

‘. .<br />

. . .<br />

,<br />

.<br />

. . -<br />

Approschcr to Improved Haaaurcment<br />

CIAUDE F. BRIDCBS, Chairman<br />

US Army Enlisted Evaluation Center<br />

Xn response to the consensus of the recomendatione of the participating<br />

agencies, the conference theme is %xraaeSng Measuring Efficiency<br />

of Evaluation Inatmentr.” This series of theoretical symposia is designed<br />

to explore ways of obtaining marked increase in the percentage of<br />

overlap between job proficiency factors and the evaluations of the current<br />

job proficiency of enlisted personnel. Special attention to innovttions<br />

and possible brenkthroughs in wtyt of increasing the covarlance between<br />

achievemant and measuring instruments Is long overdue. Concerted, lnttnsive<br />

efforts will be ntcearary if we are to make rignificant inroads<br />

into uhrt Captain Hayes aptly referred to yesterday aa “the outer apace<br />

of terting.tV Mere t’poliehing’8 activities using usual test validation<br />

and item analyoie procedures obviously art not enough.<br />

Since World War I, such polfshing activities have indeed improved<br />

tests but, as Dr. Edgerton so clearly Pointad out in his keynote address,<br />

have not made genuine changes. Perhaps the met continuously extensive<br />

research in the intervening 45 or so years has been pointed towards improving<br />

the effectiveness with which academic aptitude tests predict<br />

success in school. However,afttr all these efforts, what it the ututl<br />

correlaiion between ouch teatr and school marks? Dr. Chester Harris in<br />

the Encyclopedia of ?ducational Research states that the correlation<br />

between “intelligzce” tests and scholastic achleveoent typically fallr<br />

within the range .40 and .50. The validation studies canplated by the<br />

Evaluation and Analysis Branch of the US Army Enlisted Svaluation Center<br />

indicate that met of our recent Enlisted HOS Evaluation Tests ‘have<br />

validities in this same range. Cur attempts to adapt the most crucial<br />

aspects of the usual @‘custom-made” test developmeat procedures to a<br />

high-apted closaly-timed test production-line and to apply previous and<br />

current research findings in the measurement area to our test dlevelopxrnt<br />

8ctivities during the last six years have enabled us to improve or<br />

polish t our tests o a level approximately comparable with the typical<br />

academic aptitude tests,<br />

This looks fine. But let us consider -&at the correlation of .50<br />

actually mean2 in terms of the percentage of variance in achievement<br />

typically being predicted after 45 yetrn. Squaring .50, the upper end<br />

of the range, we get only n 25X overlap between the two varitbleo.<br />

Figure 1 presents a graphic picture of this sltuatioa. Seventy-five<br />

percent of the variance NOT touchad yet!<br />

30


w - ------- --- -.-. ___..- - I _ .__-.. .- _._.. .<br />

.<br />

Combiaetioar of two typical evaluation instruments gf.ve slightly<br />

better rasultr. The recent validity studies indicated that thin year<br />

RRC can expect correletionr in the neighborhood of those s:hown in Figure<br />

2. The Rnlirted HCS Evaluation Tear combined with ratings on the<br />

Vxxmnander’r Rvaluation <strong>Report</strong>” for mOst military occupational specialtier<br />

should yield multiple correlation (R) with peer ratings of about .57.<br />

(Peer rating8 may not be the most valid pOsBible crittrfon of all erpects<br />

of over-all job proficiency. Hovever, there is considerable eatiefectory<br />

evidence that, when obtained frcaa raters who know they are for experiment81<br />

1lp9 & peer zngs can provide practical and useful appxeele of a<br />

� ignificent portion of complete job proficiency,) We are working especially<br />

to improve the teBts for the “problem” XCS and expect even higher veluee<br />

next year. HoYtvet, the improvements still will leave not tapped about<br />

two-thirds of the factors involved in differentiating between different<br />

levslrr of job proficiency. Continued pollehing will help decrease this<br />

unmeasured variance eomevhat, but the ueuel polishing techniquer or even<br />

the addition of other carxnonly ured types of measuring inrtrumentz,<br />

conwnonly raise a multiple correlation only a few hundredths of a point<br />

at the mat.<br />

Indeed we still do have a very long vay to go. Today YB will attempt<br />

to survey some of the possible routes to be followed. Rowever, the best<br />

modes of transportation over them will remain to be determined. The<br />

general territorial area.8 which the route8 muat cover are job factors and<br />

personnel factoro--i.e., analyees of the characteristics of the job end<br />

analyses of persons performing it well vereue thore performing Ierr veil.<br />

Probably the most ixnediately fruitful of the two general areas of<br />

investigetlon involve8 the development of ways to Improve the determination<br />

of the factors in the job that dircriminata between indiv!lduels with varying<br />

levels of job proficiency, Recaure of this, and of the intorest camnon to<br />

El1 the agencfea, a special tterion on “Job Analysis for Teat Development<br />

Purpoata” hia been mt up. The rterion on l’Nonstetistica:l Criteria for<br />

Evaluating Iteaur” likewise Is pertinent to the area of job factorr.<br />

The other general area of invertigetion involves new and better ways<br />

of measuring personal factor-r of ptrronntl end of relating them<br />

appropriately to the job factora. The major typo8 of personal characteristicr<br />

for which appropriate mtesuret are nstdsd for verioua jobo might be<br />

rtructurtd AB follows:<br />

1. Job Knowledge (texonwltr of job content),<br />

2. tlentel Skills (taxonomitr of mental functione--mental manipulatione-required<br />

by the job).<br />

3. Xotor Skills (phyrical manipulation).<br />

31 .<br />

--.


-.<br />

‘. .<br />

_.”<br />

1.<br />

2”<br />

--<br />

.,;<br />

.<br />

. -:<br />

.-<br />

.z<br />

/<br />

. -.<br />

.A .<br />

.<br />

._<br />

/<br />

,<br />

,t<br />

I<br />

!<br />

I<br />

. 1.<br />

. *<br />

.;’<br />

/ . .<br />

-e__.:_I_L---.. . ._ --. -__L.- -- .YU r--- -.,-., __.. ._. _._. - -.<br />

4. Interpersonal rtlationrhipc skills (social -niPulationo) .<br />

5. Job motivation (application and effort).<br />

6. Test taking motivation,<br />

.<br />

-<br />

� � � � �<br />

.<br />

7. Interaction of personal characteristics with situatianal<br />

characteristics (Effectiveness of leadership provided;, type of<br />

management; characteristics of subordinates, peers, supervfaorr,<br />

administrators, locale, spatial conditions, etc. Why does the<br />

oamt man perform very well in some places and not in others?)<br />

8. Other noncognitive personal characteristics (attitudes, interests,<br />

level of aspiration, initiative, drive, energy, perservtrance,<br />

etc.).<br />

In the relatively ahort tims available during this conference, it<br />

will be tiposeible to effectively consider all of the major kiudr of<br />

approaches to improved mtaauremtnt of the varioua types of ,ptrsonal factorc.<br />

In order to auggtat others for later consideration, atveral potentially<br />

fruitful approaches omitted in these aaninars are mentioned. We will not<br />

discuss the atatiatical treatment of individual response data, such as<br />

beta weighted correction for errora, analysts of response p.atttrns, and<br />

scaling. Another omittad but potentially useful area would be the � nalpsta<br />

of officinl personnel record8 (use of biographical data, aptitude testa,<br />

education, age, years in aarvict, training records, honors, edvencement<br />

rate, etc.).<br />

We vi.11 be able to explore only about one idea for the other possible<br />

approaches. For example, ve can consider the utility of only the<br />

readability aspect of the ccnmnunication problem in teats, omitting semantics,<br />

use of illustrations, granxaatical conrtruction of the stem, etc. Many oE<br />

the omitted ideas involve primarily test polishing activities, but some<br />

of them ahould be remarkably effective in decreasing the unmeasured variance<br />

for a few military occupational apecialtiea. Cronbach, in his Essential6 of<br />

Psychological <strong>Testing</strong>, 1960, page 331, cite8 the result8 of an unpublished<br />

study by the Training Aida Section, Ninth Naval District, Headquarters,<br />

Great Lakes, Illinola, 1945 entitled “A Comparative Study of Verbalized<br />

and Projected Pictorial Tests in Gunnery.” He rmriaod the result8 61<br />

follows:<br />

“Training of Navy gunnrrr had been validly evaluated by<br />

acorta made in operating the guaa. As an economical aubatitute,<br />

verbal and pictorial tests were developed, Identical information<br />

vaa teated in two forma, the same quaation being asked in<br />

vorda alone or by mean8 of picture8 supplemented by words.<br />

Questions dealt with part8 of tha gun, duties of the crew,<br />

appearance of tracera when the gona was properly aimed, etc.<br />

--<br />

32<br />

. ‘.<br />

. .<br />

a<br />

�<br />

;


.<br />

. . , - .<br />

.<br />

-<br />

.<br />

2<br />

. . .<br />

.--...-. . _---- .-- --__ L.--i- . .._._” . ..- _ . .._I. __ .I__^ ---._..-___ ._.-.--.--^ . . _. - --.. .- -,<br />

.-<br />

-. __...._e._.--.I_ ,i<br />

The pictorial test had a correlatfon of .90 vith instructore’ marks<br />

based on gun operation wherea the valfdfty of the verbal test was<br />

only .62. The verbal test war in large measure a reading tes‘t; it.<br />

correlated .59 with a Navy reading test, while the picture test<br />

correlated only .26 with reading.”<br />

After these introductory remarks were prepared, the October 1964<br />

i88Ue of the American Psychologist arrived with an outstanding presentation<br />

of concept8 closely paralleling some of those we have been considering.<br />

In fact,eurpriaingly enough he used a quadrant idea to represent variance<br />

in ,comon, soqedat like this hart (F’fgure 2) that had been prepared to -<br />

-depict clearly the magnitude of O’ar challange. There is no better way to<br />

close theae introductory problem defining remarks thsn to quote the opening<br />

sentence in Dr. Melvin R. Harkr’ article entitled “How tab Build Better<br />

Theories, Test8 and Therapiecr.*’ Dr. Hark8 aptly stated:<br />

‘@The theai of this paper ia that psychological researcher8<br />

too frequently define their problem8 in ways which intrinsically<br />

. . preclude solution; that they spend their time in Bflding lilies<br />

of dubious quality rather than in determining why or how’<br />

.- promieing new blooms were blighted; further, that this<br />

.deplorable state is characteristic of theories, therapies,<br />

a n d tests.”<br />

. .-.-.....-... .-._-__,- __.I____...I____ ---- -<br />

.<br />

*<br />

.<br />

,<br />

. ,:<br />

_^._ ._-__ _ .____ _I___ -__. _ . I.._. - ._..... _I. --<br />

.<br />

.<br />

.<br />

,/<br />

$’<br />

1 iI<br />

tI<br />

i<br />

I<br />

3. :<br />

I i<br />

-


i<br />

ACHIEVEMENT i<br />

t<br />

UNMEASURED h<br />

TEST<br />

5 Y =.50<br />

r2 % Y =.25<br />

Figure 1. Percerttage of Achievement Measured by Test Hoving<br />

o Validity Coefficient of.50<br />

I<br />

67.4 9% .<br />

UNMEASURED<br />

L<br />

I<br />

&. - - - - - - w - m - l<br />

RATING SCALES - -<br />

rx2 Y<br />

P x2 y<br />

(x2)<br />

= .45<br />

= .2025<br />

P2 r = .I632<br />

- - - - - - - - -<br />

1<br />

TEST<br />

f<br />

(Xl)<br />

I<br />

I<br />

rx, Y<br />

r2x, Y<br />

PI r = .1632 I<br />

rvP<br />

= .45<br />

= .2025 ;<br />

= .20<br />

r2 Xl x2 = .04<br />

Ry*x,x2= .57<br />

R2 = .3264<br />

Figure 2. Percentage of Job Proficiency Meosurcd by<br />

1<br />

Combining Jab Mastery Test and Ratings.<br />

i<br />

I


c.<br />

/<br />

I ’<br />

.<br />

a<br />

.<br />

c<br />

.<br />

’ ,<br />

Summary Comparison of Purposes and Programs d * i<br />

:<br />

of the Ullitary Services<br />

zlr. Chairman, Rellw Conferees,<br />

E. c. JoHHSo1’I<br />

US ?rmy Enlisted Evaluation Center )<br />

As we begin to explore possible approaches to improved measurement,<br />

it will be well to review briefly, as a point of departure, the programs<br />

that arc nw in operation within the military services. You each have a<br />

detailed s-ry prepared by each service on their evaluation programs.<br />

Yn this paper I will present a Summary Comparison of the I?urposes and<br />

Programs of the Hilltary Servicer. I have prepared a char: that depicts<br />

thi 8 canparison. his chart is listed as Appendix 1 at the end of this<br />

paper. All the military services use a paper-and-pencil multiple choice<br />

type test as the basic inrtrument of their evaluation progrsm. This<br />

te8t is supplemented in some cases by a performance test. These tests<br />

mre very similar, vith the exception of the Marine Corps, in that they<br />

are designed to mcamre the job knwledga required for satisfactory performance<br />

in the militury specialty that IS being evaluated. Each military<br />

mpecialty has a separate test for each subdivision of the specialty.<br />

These subdivisions are either pay grades or skill levels. The number<br />

of questiono in these tests varies from 65 to 150. he NC0 tests include<br />

coverage of general military and f.lpervisory abilities as vell as the job<br />

specialty.<br />

.ARmPRoGRAH<br />

The Army uses two inotruments in its evaluation progrsm. An M0.S<br />

evaluation test for each skill level and a rating form, Couxnanders<br />

Evaluation <strong>Report</strong> (CER). The CER Is a rating form ueed to evaluate a<br />

noldier’s performance characteristics in a specific HOS at an ertablished<br />

level of skfll. One rating is accanplished by the soldier’r immediate<br />

superfor and another by the immediate superior of the rater, The CER’s<br />

include scales for rating the individual’s cooperativeness, reliability,<br />

job performance, and other factors.<br />

After the tests are administered Army wide, the MOS evaluation test<br />

unswer cards and the canpleted CER’s are forwarded to the USAEEC. The<br />

two Instruments are scored at the Center by canputer and aI canposite<br />

evaluation score is obtained. An HOS Evaluation Data <strong>Report</strong> is prepared<br />

for each individual teuted. This report contains the individual’s<br />

Evaluation Score and a profile showing his standing in each subjectmatter<br />

area of the test. The unit coaxmander receives a copy of the<br />

individual’s EDR which he reviews and then forwards to the individual<br />

soldier. Thus the individual and his unit commander is made aware of<br />

the subject-matter areas’ in which the soldier stands “high” aa well as<br />

those areas in which he needs to improve.<br />

.<br />

.<br />

.<br />

. .<br />

._ .-- --.. - _--.__ _ ___ _ _-_ -___ -.-__,._ XI ..__w_._. ---I_-- --.. _.<br />

_ _--__ -_ -. ---. - --<br />

.<br />

i<br />

:<br />

i


.<br />

, -.<br />

.._..<br />

.-<br />

---- _..-.__ 1-1 .-.__ - ___... _<br />

Thir FSX Evaluation Score and the Profile Data <strong>Report</strong> i-8 then ured<br />

a8 a baris to:<br />

Source of Itmu<br />

1. Avard Pro Pay<br />

2 . nos Verification<br />

(Primary and Secondary HOS)<br />

3. Ronmtiou Qualification Score<br />

(U8ed by camanders on m optional bario)<br />

4. Pay Grade end MOS Determination<br />

(Officer reverting to EK)<br />

5. Identify Training Reeds<br />

(Both for the individual end the unit)<br />

The Army depend8 upon Army service school8 and other inrtallatioas<br />

to kite tbair te8t itaac. Host of the tlmo the item writer8 are lnrtructorr<br />

in the MO.5 qualifying coume. They are civilianr, enlirted men, or<br />

officer8 who are arrfgned thir duty in addition to their regular inetructoz<br />

re8ponribilitie8.<br />

The Nuvy u8a8 a 8sparrte examination for each petty officer pay<br />

grade of all Navy rating8 with a few exctptionr. That exuainationr are<br />

� dmlnlettrtd by t.xrmin?.ng boards vhtttvar naval parronntl are ttationtd.<br />

The completed tn8vtr card8 ure scored by the Naval Exuainlng Center.<br />

Pars/fail cattgorltr art 8et up for each exmnination bared upon a multitude<br />

of factor8 much as netdr of the rtrvict at large, budgetary problase, and<br />

above all, the dtgrtt of qualification and performance of each candidate.<br />

Each candidate ha8 to par8 the exmainatlon before the other factor8<br />

� re conridered. Final � dvancanant, � sauming the candidate has fulfilled<br />

all prtrrqui8itt rtquirarentr and has attalned a par*ing #core on the<br />

examinatfon, 18 made on the baair of relative rtanding on a final cwpoaitt<br />

acore. Thir canposits score includes there five factors with cueximut~<br />

valutr U8 follour:<br />

.-.._<br />

.<br />

. .<br />

.<br />

.<br />

_. . _ ., _- .._ _ _ - .<br />

.<br />

.<br />

36


--<br />

/-’<br />

1. Bxamfnatfoti Grade<br />

2. Performance Factor<br />

3. Length of Strvicc<br />

4, Time in Rate<br />

5. Naber of Awards<br />

Maxiaum Composite Score<br />

If the candidate ccnzpcter In an occupation where there am more<br />

vacancies than qualified ptrsonncl to fill them,then the exanination<br />

acore qualifies him for advancement.<br />

This is the procedure used for advancement to pay grader E-4<br />

through E-7 to fill Navy wide vacanqies. Selection of personnel for<br />

pay grade8 E-8 and E-9 to accaaplfrhed by a selection board convened<br />

in the Navy Department, vhich rtviewr exuninatfon result@ and fndividual<br />

record*.<br />

Source of Items<br />

The Navy’s item writers are aseigned to the Naval1 Examining Center<br />

for a regular tour of duty.<br />

TEE AIR FORCE PRCKZUX<br />

The Air Forct uses three rpccialty knowledge tests for each Air<br />

Force Specialty. They are 3 (#emi-skilled), 5 (skilled), and 7 (advanced)<br />

levelr. There epecialty knovledge terto are adminirtared to Air Force<br />

ptrronnel world-wide. After the 5 and 7 level tests have been given, the<br />

teat ansver cards are sent to the Personnel Research lAboratory where<br />

they are rcored. The paes/fail rates for these test8 are determfntd<br />

in accordance wfth criteria ertablinhed by Htadquarterr, US Air Force.<br />

Scores on rpecialty knowltdge ttata are reported In pcrctntile form and<br />

do not indicate the actual nunber of questions answered correctly. This<br />

score ie used to identify those airmen who posrerr sufficient knowledge<br />

about their jobs to be considered for up-grading to a hlghtr level job.<br />

Thie ecore ir then used vith &her factors to areers the overall compettnce<br />

of airmen for the avard of an Air Force Specialty.<br />

Source of Items<br />

The Air Force pl;:er ftr subject matter apecialfets on TD’f to the<br />

Personnel Research Isboratory. These subject matter specialists are<br />

senior NCO’s in tht Air Force Specialty. for which tests are being developed.<br />

They develop all the test items for the three tests in their specialty<br />

while they are on this TDY.<br />

37


’<br />

;;<br />

. ‘-_<br />

. w<br />

__ .,-_. . __-__._~.---~~x~~~~~c~ ---- - --_.._ . ._ - __ --.- --.. ---<br />

Tbb c’c -t @ad’s progrrs barically porollelr that ‘of t’he Navy<br />

80 I will not go into their program in detail.<br />

\<br />

Tffe KARXNE CORPS PROCRAH<br />

The Marine Cor~8 utiliner revere1 test lnrtnments in their evelu8tion<br />

program.<br />

The tart that ir 8-t caqmr8bla in u8e to the evaluation te8t8<br />

urbd by the other rbrvicb8 18 their Caneral t4ilitary Subject Tart (c;MsT)<br />

Thiv telt 18 ured to determine the eligibility of enllrted men for promotton.<br />

It 18 adminirttred to eligible corporal8 through rtaff rerguntr.<br />

(E-4 through E-6). It differ8 frau the other tt8ts dl8&88ed in that it<br />

doe8 not cover the militcry rpeciolty or occupation of the individual<br />

but h88 a broad rcopt ccncanpa881ng thO8e rubjectr considered inderpenreble<br />

for Mariner in a canbat area. Tha8e rubjectr include tactical subject8 at<br />

8qu&d and platoon level, equipment and uniforml, field 8anitAtiOn, fir8t<br />

aid, etc. Thin tt8t 18 given three time8 trch yur, the Mower card8 art<br />

8ent to the Marine Corpr Inrtitute where they are gmded, and the rerulto<br />

are foxwarded to Promotion Branch, He8.tquarter8, Marine Corpr end the unit<br />

cmnranderr .<br />

A Urine rm8t pa88 tha CZGT before he i8 conridered for prarotim.<br />

The other factor conridered ir hi8 promotion 18 hi8 proficiency in ha8<br />

technical 8QOCiUltv. Thi8 IO determined by UImU8 of fitness report8 rubmitted<br />

by hi8 immmdiate ruperior over a period of year8.<br />

A8 8 matter of gtntrwl interbrt thb HArime cOrQ8 h88 threa other<br />

tbltl.<br />

1. The Geucrql <strong>Military</strong> Subject Proficiency Evaluation Tert j@lSPET)-<br />

Thi8 tart cover8 genernl military rubjtctr in more detail than the<br />

Q4ST and ir urbd by the unit commander to identify the training need8<br />

tnd to judge thb bffeCtiVtie88 of hi8 training progrun. It i8 provided<br />

to thb wilt c- dsr by the Marine Corpr Inotitute. It ir provided<br />

8ohly for the locel caruaander. No report m tb8t rb8Ult8 arb made to<br />

highbr hudqtmrtbro.<br />

2. Inrpector General Te8t (IGT)-<br />

Thir tart 18 an bxtract of the GKSPET with limited rrrnpling. It<br />

i8 deeigned for ~80 by the XC on hi8 Mnual in8pbCtiOII. here tart8 are<br />

‘gradad by the Marine Corps Inrtitutb tnd rtrultr arb nuilbd to thb XC<br />

and unit in8QtCtbd. The r"QOrt rbflbCt8 Unit QWfOlDWCb and CCmtQarb8<br />

the unit with other typo unit8 end thb Marine Corpr norm.<br />

38<br />

-- _-__. . -- _.._._- --- - - --- _.._...- -_- _--. - .-<br />

.<br />

. .<br />

.<br />

,<br />

-.<br />

c<br />

.<br />

.<br />

.<br />

.<br />

:<br />

- _.. . _ . . . r . ( .. -'.. . - -. * _ .._m __ ,__ __._ _ -.--- ---- _.- - . -I .--<br />

.<br />

.<br />

.-<br />

.


: .<br />

A<br />

.<br />

.<br />

.<br />

3. Officarr’ Administrmtivs Subjects ExamlnatFon (QASI?~<br />

The purpose is tolwtivaCe avery company grade officer to faxitfarirrtlm<br />

with a varfrty of edxafnirtrativo typo eubfectt. Every officw<br />

muat t&a thm aroadnctfon, re a Lieutenant and again ae (II Captain.<br />

Soccurful cafspleticm of ME is not adatory for promotion but arch<br />

officrr baa to continue to take the cxmdnatla tech peer until ha<br />

ruccesrfi~lly parrrr it. The test itsar are eelected by CL board of fiald<br />

grade officero and acre&led into a te8t at WI. 3-m temt8 are grsded<br />

by m and 8 r8port 18 8ubmitted to\XArine Carp He8dqUarter8 which in<br />

turn publirhrr the rerults and direct8 an rpproprints entry in th8<br />

OffiC8r'8 jACk8t.<br />

lor thir 8~mtp08i.um, tvo pertinent conclu8ioa8 cw b8 mad8 -- 1.<br />

Althargh th8re are rlgalficant difference8 in th888 progr8m8, th8r8<br />

oboiour~y 16 con8td8rable CanauMlfty fa the U8es of th8 tcrto. 2.<br />

Purth8nmre, th8 baelc goal of th8 programs of all tbs rervfcco c8n b8<br />

srpt8408d, in a gtnrr81 v8y, ~8 being ured t0 provld8 mU8ure8 Of thr<br />

extent to uhich wincer have nmtsrad the rignific8nt aEpact8 of<br />

thrir milltorp 8peCi&lty.<br />

! : 39


I<br />

.a ,<br />

I<br />

’ !<br />

; * .<br />

.<br />

. . .<br />

.<br />

,<br />

Am!t-<br />

NAYY<br />

AIR FOt Cl!<br />

COAST GJAxD<br />

zi<br />

TEST<br />

1<br />

Mm!&<br />

‘IEST<br />

.<br />

. * .<br />

AWARD PRO-PAY<br />

WOS YERIP ICATIOW<br />

FFuxYrIm QuALxFICATmN !SJBJECT-tL4TTKR SPZCIALISTS<br />

EVALWTICH SCORE SC~RR AT SERVW? SCHoOts AND<br />

PAY GRADE 6 tfOS<br />

Dl!mmma’L?rn<br />

IDEHTXFY TRAIMING NEEDS<br />

CYI’fiER AREt INSTALUTICNS<br />

PEW-E FACTOR<br />

SUBJECT-IUTI’RR SPECIALISTS<br />

x.mcTx OF SERYICL PRm(xp Assn?zED To FUVAL<br />

TIXKINRATE<br />

NUMKR OF AWARDSIIt<br />

EXAtUMI!G CFXTER<br />

TRST<br />

UI’RER FAcfCRS<br />

TZST<br />

PERF-CE PACKtR<br />

mm OF SKRYfcE<br />

-liiEZYXEARD,J<br />

SUBJECT-HATTKR SPECIALISTS<br />

AWARD OF M sPEcIALxlp TDY TO 6570th WRSmL<br />

RKSEARca IABCRATORY<br />

PRmoN SzmcT-#hTrRR SPECIALX~S<br />

PIMXWTIOR @MST)<br />

TEST IDEmIEY TRAINING<br />

NEEDS (CXSPE’T) SuBrn-PATrKR sPEcr.hLxsTs<br />

CYl’ER FACTORS<br />

nvsPEcTxoN (ET)<br />

tiOTIVATIfXl (OASR)<br />

. . . c<br />

.\ -_-_--<br />

, .<br />

. *<br />

I<br />

. .<br />

I


.<br />

.<br />

.<br />

Conctpta for Exploration in<br />

Prof iciepcy tieaaureraent<br />

EfICHAEL A. ZACCARIA<br />

Lackland Air Force Base, Texas<br />

XARVIN KAS’<br />

Randolph Air Force Base, Texan<br />

Roth the keynote speaker and the cbairam of the tboretical aympoaim<br />

on approaches to improve measuremat have asked that ve develop C~ccpte<br />

for exploration in proficiency mBaaurcment. Our theoretical sp~~osi~m<br />

chairman baa indicated that vc can gain very little by merely polishing<br />

up our present meaauremnt techniques. Thus, any approach to a real<br />

improvement in proficiency measureuxmt muat be a radical departure, at<br />

leaat in aow vaya to the conventional concepts presently applied in<br />

this area. The traditional appmach has atereotyped our thlnklng to such<br />

I an extent that ve are primarily concerned vith marely ru Lning and velidating<br />

proficiency measurement devices. It might be vell for ua to bc:!r.<br />

this discussion by reappraising aoaz of the concept8 that have ‘


..- .<br />

-*- .,-<br />

‘. _<br />

.<br />

When the matn objective is to use a umaoure to select or screen<br />

individual8 for a .program, there ia no question concerning the value<br />

of a norm-referenced syatcm. On tha other hand, the question arises<br />

ar to whether we should group achievement measures in order to con!4 up<br />

with one score.<br />

()ur program chairwa has stated quite tuccinctly that in order to<br />

mke real progress in thir realm of proficiency evaluation, id, .&et<br />

begin with the job. We are definitely in agreement vfr.h this comaznt<br />

end wish to present some concept8 of how this r,:$c be done. Sa<br />

individuals have approached thte task by :L+ use of 6upervi60zy ratings<br />

as an ultimate criterion on the ow ‘jtrd and the development of tests of<br />

proficiency on the other. ‘31 approach has been one of intercorrelating<br />

the teut batteries ar.d crl:eria thuo deriving optianm prediction weights<br />

for each test.<br />

Thrs courm of action ray vell remind us of the story of a team of<br />

prychologirtr baLrg called upon to develop prediction devicas for a large<br />

company. After considerable work on these predictore, thio group wa8 &la<br />

to predict quite highly with a large battery of temts the criterion which<br />

war based on rupervirory ratings. The paychologistr, being dfasatirfied<br />

vith marely being able to forecast there ratings with naariy perfect<br />

accuracy, decided to do mme further research and to factor unalyze the<br />

criterion and predictor variables: Lo and behold, the analytlia being<br />

complete, they found that they were Pearuring job saniority.<br />

Validity of perfownce amaouree riced not ba related to another<br />

criterion. These could be ultimate criteria in and of themselves. Let<br />

UP ruppose that we wish to ueaaure the performnce of secratariee. The<br />

first thing we should do ir etart vith the .job. We should enumerate the<br />

taoks that she doar in terma of product-lfke performr.ice. A secretary,<br />

for example, may do the following: type letters or finished copy fros<br />

draf te , take dictation, file correa~ndance, #elect mterialr from filer,<br />

receive vfstoro, and answer inqutrfer over the telephone. From this<br />

brief description of the recretarial dutleb, we can quite reudily develop<br />

uome of the following criterion perfonmnce measure#: We can wasure her<br />

speed and accuracy of typing, her rpeed and accuracy of taking shorthand,<br />

and her ability to transcribe her shorthand notes, her abill y to sppropriately<br />

file corrasporxience and to find mterials filed. Her ability<br />

to anmrcr qumfer on ths telephorm may be somevhat mre difficult. On<br />

the basin of each score that rhe maken, we can clarrffy her in terma of<br />

ratisfactoty. unratiefactory, end outstanding.<br />

The criterion for satirfsctory on typing, for example, mfght be an<br />

average of 40 wordr per minute on a cartrin specified copy vith no more<br />

than two error43 per page for a two-page amnuscript. Outstanding on thir<br />

charactari8tic mfght imrolve 60 words per minute with no mre than one<br />

error on a two page manuscript. Sfmllar stemjards could be worked out<br />

for each of the other criteria.<br />

. . . - _ _<br />

.<br />

.<br />

_ _.<br />

.<br />

_ .-<br />

42 .<br />

I<br />

-.<br />

c<br />

i<br />

.<br />

. . .<br />

.<br />

.<br />

:<br />

* .- ‘, .’<br />

.+<br />

. - . - ,.,- i... .-_.. -...<br />

__ _ _. _. - - - .- . ._ -_- . _-^_-dA- _______...._ - ^-.- .--. - -.<br />

. _<br />

,<br />

.<br />

.


Here; UC are trying to make acveral pointa. k’s ten define and<br />

develop criteria in term of certain relevant on-the-fob bchsviora,<br />

Each of the&e different types of behaviors coo be rrreaaured separately,<br />

and they would be made more meaningful if they-were aeaeured in terms<br />

of relevant requfrements. It Is quite clear that 8oa.a secretarial jobs<br />

may require different: degrees ‘of skills. Individual skills could more<br />

readily be matched to required job akills. There is no need to coc&ine<br />

all of these acorea into one score and then further putting this score<br />

into a rtandard score or percentile conversion. This obscures the data.<br />

Purthermrc, there is no need to correlate each of these scores<br />

or obtain a weight for these scores by optimlly correlating these<br />

againat auptwisory ratings. After all, it could lfkely be that WE<br />

throw out the baby with the bath because the auperviaory rating may<br />

not be baaed on any of the charecterI8tIca that are relevant to actual<br />

job � ucceaa . We have all at least heard of cases in which aecretarlaa<br />

were rated outstanding by the boas even though he has never deen them<br />

do meny of the taaka typically required of a secretary.<br />

Teat Reliabil1tp and .Accuracx<br />

For the purpose of this presentation, we‘have discerned three types<br />

of teat reliability. The first type deals with the cqufvalence of two<br />

or more form of a teat. The second concerm itself c:Lth the stability<br />

of the seam teat given at two djfferent timor. Youmgenelty or sameness<br />

throughout the teat ir the third kfnd. There 1: w question as to the<br />

importance of each of these types of reliabflfty .Y connection with<br />

aptitude and peraonaiity testing. A question does .clae, however, as to<br />

the appropriatmeaa of atabllity and hamogtneity coe...cienta In achievemant<br />

testing. There ,La no doubt a8 to the importance of equivalence of<br />

form for a particular achievement teat. The present author8 raise a<br />

question on how one determines the equivalence of hro fonaa in a training<br />

situation.<br />

In the ideal training situation any individual or group of ‘ndivduala<br />

would theoretically score zero on any test at the beginning of training<br />

and LOO percent at the end of training. While it la true that few, if<br />

any, ideal situations of this sort ever alat, it is important to take<br />

the ideal situation for conaidcration. In an ideal rituation such as<br />

aamtioned � bwe, If we used end-of-cmrae achievement teat as a measure<br />

end used the traditional coefficient of correlation as the tool to<br />

determine the degree of equivalence of two forma of a teat, we would<br />

find that the correlation coefficient is not very high. E%en in a amwhat<br />

less id-al situation vhere ve have acorta bunchlzg up at the low<br />

end during pretesting and bunching up at the high end at the end of the<br />

course, we would not obtain a very high reliability coefficient by computing<br />

it based on only poatttat acores.<br />

43<br />

I


. .<br />

-<br />

/:<br />

i<br />

.<br />

. --_<br />

- _ _. _<br />

.<br />

-. -. .-- . . ..-----...-;-.---.-<br />

_-...II -.--.- -_.. _.<br />

our proposal is, thus, that if we muet hnve a correlation coefficient<br />

:o indicate tha equivolansa of .two forw, that data be accumulated<br />

at the cad of vatloucr portiona of P muroe of ins*.ructioa. ‘fw0 forlao o f<br />

a test could be adAniatcrad :J 80~x3 indi~iduid a8 a pretct6t, t0 othar<br />

Individuals ,durfng the courcc of itwtruction,and to othere at the oad of<br />

. the COUTC~. In this mannat, a correlation coefficient would shw the<br />

true relaL:on.ship between two forms of a test rind would thca r,ot be low<br />

due to restriction of a tent and would thus not be low due to restriction<br />

of range when students arc vcll taught.<br />

Another posrfble manure of rslisbility of a teat would be in terms<br />

of accuracy vith which it vaasures that whfch it is suppoead to umiure.<br />

This ir closc’ly akfn to relevancy azA validity of zmmuremmt and vi11<br />

be discussed in the n)rt rcctioa.<br />

Tart Validity and ValKdation<br />

There are a aumber of crroneoua ccnceptr concerning achievmmznt taatfng<br />

which rhould be clarified. when we speak of t~hievment or proficiency<br />

tedting, we are, in emenee, dircusrfng a criterion. If thfs bo the ease.<br />

vhy do vc have to wonder whether our teat predicts job nucccss? Wfliltl i t<br />

is true that there-are more than just job comgomnta ta the sumesaful<br />

psrfmnce Of 8 job, if WC can 8t loAnt ACcurat6ly nremrure the VariOUII<br />

cmpments of a job, we will have gone a loa vay toward thu actual<br />

devclopnt of a crfterior.<br />

There are those who feel that tsmy echlcvczmnt or pcrfonmncc teat8<br />

must correlate highly with an ultinrte “criterion of job 8uccem 08<br />

dctsradned by supe~~sory ratings.” We have already -de our point<br />

concerning this aspect. If we have szaeured a seczetary’a proficiencies<br />

in trF& of speed aud quafftF of typing, .shorthand, and fLling, we have<br />

already gona a long way towards umaruring her total job performance,<br />

although we nay hma naglectcd to rrrtaaure her initiative and mtivation.<br />

There tvo latter F?;jlzits, we vi11 agree, can better be measured by supervirory<br />

ratinga. We see no need, on the other hand, to validate the other<br />

measures by correlating them wCth � upenrisory evaluation of perforrcancc.<br />

, Ue do aot thfnk that the traditional practice of obtaining item<br />

total test correlations on achievement tests in very worthwhile, Sam?<br />

test developers seem to forget that this practice leads to homogenlafng<br />

the test, A practice which is generally acceptable in aptitude of<br />

parsoumlity test development. In the case of achievement testing, hwever,<br />

this practice has very limited usefulness. He feel that item<br />

validation could better be achieved by admiairtrat


‘a<br />

: .<br />

-..<br />

. .<br />

1 .<br />

. .<br />

. .<br />

_. . . .<br />

‘.<br />

.-<br />

. I<br />

---__ _ _,___ - --_- -- . _-__ i_ -___-___ .__ __._ _. . _ ‘_ -. -_ - - _-.. _-..- .-<br />

.*<br />

;<br />

Pf we aauot valldate echievcmcnt and performance teoto ngajnrt<br />

on-the-job criteria, certain conditions muat be preoent. Part of the<br />

on-the-job criteria murt. include oimulated or real taoko or job ample<br />

.s<br />

teoto. Each of there job criteria ohould bt measured and ‘deoignatcd<br />

reparately . The relevant rchool or trairiing criteria should � l8o be<br />

measured in a 8imilar manner and relationrhipa should be eotabliohad.<br />

If correlation8 are desired, they can be obtained by correlating the<br />

variour relevmit 8cor’co of not only those individual8 who 8UCCeS8fUlLy<br />

complete a cour8e of inotruction, but a180 of those who do not OuCCtEQfully<br />

complete, or vho go on the job directly without any training.<br />

:<br />

We feel that there ruggcstionr will open new avenue8 for axploratioa<br />

in proficiancy an’d � chieveznmt evaluation. i<br />

.<br />

_<br />

Referenccr i .<br />

Prederick8en. N. Proficiency te5t8 for training evaluation. In Glaoer,<br />

R. (ed.), Psychological research & training and educatCon.<br />

Pittaburgh: Univerrity of Pitteburgh Prerr, 1961.<br />

Glaoer, R. Instructional technology and the mcaourement of learnfng !<br />

ou tCOr888. Amer. PlychOlOgi8t. 1963, 2, 519-521. i<br />

Glarer, R. 6 Rlauo, D. J. Proficiency meaaurfmmt; � 88e88ing human<br />

performcncs. I n Gape, R. R. end Gtherr, Psychological principlea<br />

& ryotcm development. New York: Iiolt, Rinehart, & Winrton, 1962.<br />

- -<br />

Nedelrky, L. Abrolute grading otandardr for objective tertr. e.<br />

Poychol. Measot., 1954. lb, 3-19.<br />

Thorndike, R. L. 6 Hagen, E. Haaourement and evaluation & p8ychology<br />

fi education. New York: Wiley, 19557<br />

Tyler, R. W. Achievement tc-!ing and curriculum conotruction. In<br />

Williamoon, R. G. (ed.), Trcndo - in - otudent pcrronnele.<br />

Univerrity of Hfnnsoota Prero. 1949.<br />

Zaccertq, H. A. h Oloen, J. Reapprairal of Achievement Raaourer<br />

USAF i.:-tr. J., 1953. 1 73-75.<br />

,’<br />

4 5 i :<br />

‘?<br />

!<br />

!<br />

1<br />

� �<br />

i<br />

t<br />

x<br />

i<br />

i j<br />

j<br />

I<br />

.I<br />

:<br />

I<br />

I P---<br />

�<br />

i<br />

___-.-..<br />

.<br />

.<br />

, . .<br />

_ - .- i.... -__ _._ “_^_~ . .__ -._ _. _.^.____. “_. _. - . _.-- . ..- .___a __.__ _.-.- . .._..--.-_-. -.- “.-.^..I-- ---- -.A.- _ _._--_. _-_.._ :..,.-, -._- --.-. -.---. ._,-_._ - . _______*__ _----- . _ - . - -_--.-_ - -<br />

. t<br />

i<br />

-III_L-- i


--<br />

.r<br />

. .<br />

“\.<br />

.._-.<br />

-.<br />

.- -__.<br />

,<br />

.<br />

-.<br />

,<br />

.<br />

.<br />

.<br />

.<br />

. --_ ._<br />

-----. - _ _ ._-._ -.._ ._ _ ^ .- . -- ._. _<br />

Item ~nalyrie Information for Test Revision<br />

VERN w. URRY<br />

US Arwy Enlisted Evaluation Center<br />

The item aMly8is coding procedures used at the US Army Enlisted<br />

Evaluation Center represent an application of Cullikren’e item &leCtion<br />

techniques to MOS Evaluation Test redevelopment. These techniques were<br />

advanced by Culliksen (1950) in his “Theory of Mental Tests.” They call<br />

for the use in item anrrlyeis of the point-biaerial correlation coefficient.<br />

Ws are currently using thir statistic for that purpose.<br />

XtCcP analysis codes are provided to indicate appropriate actions to<br />

be taken by item writers in regard to individual HOS evaluation test items.<br />

While a full explanation of the 8 Codes used vi11 not be given, the present<br />

paper will discuer the statistlcal bases for theme action code8 vhich have<br />

to do vith the selection of acceptable item or the revision of items<br />

needing minor changer. The advantage8 fn improved measurement3 to be<br />

gained by the implementation of these particular action code8 will al80<br />

be dircusscd.<br />

Basically, item analysis has am its purpose the development of a<br />

test vith a given set of statistical chararteristica either through<br />

item eelectioa or item revisfoa. For a lOO-item HOS Evaluation Test,<br />

the deeired set of statistical characterietice vould include a mean of<br />

62.5, a standard devtatioa of 12.5, and a reliability of .80 a6 a mFaimum.<br />

The desired mean equally divides the range from a chance score of 25<br />

to a mxinarm or perfect score of 100, and the derlred 1taadard deviation<br />

io determined by dividing 6 into the above range, since 3 atanda:d<br />

deviations on c ither side of the mean cmtaine, for all practical purposes,<br />

a normal dictribution. The chance scora ir a function of the fact that<br />

each l4m Evaluation Tec;t Item contains 4 plausible alterridtives; hence,<br />

if OM were guessing 11’4 or 25 of the item would, on the average, be<br />

msvered correctly.<br />

The it- analysis action codas to be useJ by iteta writers take<br />

into account the dcrlred set of statistical characteristica for NO!3<br />

Evaluation Tests and the item statistics required to cbtafn HOS Evaluation<br />

Tart8 vith these atatirtical characttrirtics. Item rtatisticr relate to<br />

the statistical characterirticr of a teat in the following annner<br />

(Culliksen, 1951):<br />

.<br />

.<br />

46 ’<br />

--. . __. . ..-- - - -_-_--<br />

.<br />

I<br />

e .<br />

.<br />

.<br />

.<br />

.<br />

.<br />

.<br />

:<br />

.<br />

_.. ,.<br />

‘-. .t .<br />

^I. .^_<br />

.<br />

.._.._ ^-<br />

. ..-<br />

.-.-<br />

- -- -- -. ---<br />

.<br />

#<br />

.<br />

.<br />

!


f<br />

-.<br />

‘Ibotertfaoul,X:-Cp I<br />

.-<br />

.<br />

._.<br />

.<br />

(Formula 1)<br />

Where: pi, or item p-values, la the<br />

proportion of individuals<br />

anawaring the item cor-<br />

, ractly<br />

The ‘teat � tan&rd deviation, s, - Critai (Foraul* 2)<br />

The teat reliability, rm =<br />

Uhere: rit8i la the point-biearial<br />

item-teat correlation<br />

multipled by the itar<br />

etandord dsviatior~<br />

‘1 ‘JPi (l-pi)<br />

(Formula 3)<br />

Where: I: lo the numbsr of iteqa in<br />

tho toot,<br />

� � 1 the itaa variance<br />

si<br />

‘I - Pi(l-Pi)*<br />

tits1 la the point-biaorial<br />

itaa-tect correlation<br />

nultipliod by the itcap<br />

standard devir tfon<br />

In this ragcrd, ony itt-a action coding proceduroa that moues from<br />

an ito analyaia should taka into account the end produce, or HOS tvaluation<br />

Teat, which vould result, given the selection or revlaion of itwa<br />

to k included in a rodevelopod teot.<br />

In keeping vith a deaired toot mean of 62.5, all itrn p-valuoa � hould<br />

������� � rouod .325. For a doaired teat atonderd deviation, I:- p-values<br />

uhould also rango around a mdian value, I.o., a8 nmr .50 &I la practicable<br />

la view of the dealred mean, rlnce al - pi( -pl) and lo at a maximum vhm<br />

pi - .50.<br />

Ap-?<br />

The mlacted or acceptable items � o coded in the present procedure<br />

with codes OK or M (mia1nslly acceptable). Itana to be rovlmed, Code<br />

RV:, reach the HA level for item-teat correlations but either dirtractorteat<br />

cotrelations or distractor proportions indicate they need minor<br />

modification to be in consonance with tho � tatiatical characterlotica d<br />

doaired in tt0S Eveluetion Teata.<br />

, i I<br />

. .<br />

i<br />

I


. .<br />

.<br />

_. .- --\<br />

..-._-_- . .._<br />

;--. . :<br />

.: ‘.<br />

. .<br />

‘,<br />

(- . .<br />

.<br />

.<br />

. .<br />

I<br />

.<br />

‘._ *<br />

-. 2’. -.‘.<br />

. s<br />

_ _. -.- : .--..- .: -.._..- -- ..--;-A---- -___ ._<br />

.<br />

1<br />

I<br />

Distractor-test corralstions indicate revision is ctcessarp if they<br />

exceed the correlation set at the HA level. Distractor proportions are<br />

indicated for rtvirion if they fall outside a range fro66 a proportion<br />

of .OS to a proportion of .25. A minimum proportion of .OS wnd elected<br />

to provide that scm6 plausiboility exist in all distractore in order thut<br />

a theoretical chaacS6 scort.of 25 is teuablt. A msxinarn proportion of<br />

.25 wa6 btltcted becauat (1) all rc6pons6s, distractox and correct<br />

*<br />

altemativts, would have s proportion of .25, if examinaao had no icnm- .<br />

ledge regarding the item and (2) when the.p-values of correct alter- ’<br />

native6 approach .25, chance rc6ponses tend to lovar item-tsot correlation6<br />

(Gullford, 1956, p. 437). .<br />

. *<br />

. .<br />

Applied in conjunction, the above limits provide: (1) that :tem<br />

p-values lrvy range ‘bstween .25 and .85 which ie rearonsble in view of<br />

the desired man; (2) that the resultant test mssu is near the desired<br />

valus; (3) that wximm item standard deviation8 art obtained rince<br />

item p-values am msintainsd at andiaa values; and (4) that item p-valuer<br />

trt within ranpr wherein item-tart correlations art higher since extrexe<br />

or nonmedian p-valu66 for items tend to reduce item-test correlations.<br />

To iIlu6trata how Ponda 2 works in regard to the item coding<br />

proctdurtt, we can take th? txmple of ittw coded CX and HA:<br />

Given 100 item6 with item-ttrt corrtlacions of .30 (the tX level<br />

for larger NCS 8+6ples) and item standard deviation6 of4-9<br />

.48<br />

the test standard dfmfation would be:<br />

Ix - 100(.30)(.48) - 14.4<br />

With 100 item having itmn ta6t correlationa of .20 (the minimslly<br />

for larger HGS ramplea) aud i-em standard deviation6 of<br />

.48, the test standard deviation would be:<br />

sx - 100 (.20)(.48) - 9 . 6<br />

’ ; With item 6tleCttd or revised to txcttd there item-test correlation6 and<br />

to � pprosch these item standard dsvistions it in apparent that a daairtd<br />

test standard deviation is tmdt more posriblt.<br />

!<br />

‘lb illustrate how ?oxw~la 3 worka in relation to tbt item coding<br />

procrdurtr,wt can use 100 items with the average variance of (.625)<br />

(.357) - .23 and a test standard dwiation of 12.5. The miniatum tt6t<br />

reliabflity in thir cast would be:<br />

rxx - loo<br />

99<br />

l-.=&-.86<br />

In this msnner adequate reliability tends to be iasursd by the item coding<br />

procedure6 .<br />

48<br />

.<br />

.<br />

.<br />

i


-..<br />

:.. . ,<br />

/<br />

_ :.<br />

. .A.<br />

.<br />

.<br />

-<br />

---. *<br />

I<br />

.<br />

;<br />

. -<br />

i<br />

, -.<br />

.<br />

.<br />

Since &zm statistics ralata to the statistical charactcrtstfcs<br />

of tests as ,indiceted fin tb above formulas, it is further possible<br />

for test spclaliats to exercise more control wet thase aspects. This<br />

furth3t control fs possibla becPus8 each stetistfcal charectaristic of<br />

a redeveloped test can I32 estimated from Ftem ztatfsttcs which ~111 %e<br />

provided on each 140s Rvaluation Item Card. The sumatioa of these .Ltcm<br />

statistics and entry Znto the proper equations for the eatkstion of<br />

aech of the test statistics is thereby fccillated. Items to be included<br />

in a redeveloped tent could then bo varied in order to more closaly<br />

approach tha desired set of statfatical charactcrlstics.<br />

la sunstmry, action coding procedures tsnd to channel efforts in<br />

item selection and item revisfcn touardr HOS EvalustFm Tests of dosired<br />

characteristics; and access to pertinent item statistics on the part of<br />

test specfallsta vFl1 prwlde for a more accurate approzcinution of the<br />

desired set of statistical charactertstics when redeveloped tests are<br />

subsequently used tn evaluation tasting.<br />

Reference<br />

Cullikaen, H., ‘fieor]! ef F?ental T& Xew ‘York: Wiley, 1958.<br />

.I -<br />

49<br />

‘.<br />

!<<br />

,!<br />

�� �<br />

.:


. -. ‘,‘,.. . . ‘-’ _- _.<br />

1. *’ * _. L-.--- -’ ._--- c ._.__ ..’ -:.‘:.--2 .. - --.---_-.2.e__--P ,-*<br />

.--m--L. --. __.. -_ .- ..._I_.___ _______<br />

I<br />

I;.;. : .i<br />

, :<br />

I’<br />

: .-<br />

,<br />

. . ,<br />

L<br />

/;<br />

s’<br />

;.<br />

, . L<br />

,.-<br />

,<br />

Item Validity Informetion ee a<br />

Beeie for Teet Rtvisim.<br />

VERN W. URRY and EDWXN C . S’RIIKEY<br />

US Amy Enlisted Evaluation Center<br />

. -<br />

Xort of ue would pro’bmbly agree thet ~concepte such 88 tee; reli- .<br />

ability and content validity are important chuecteriotice of teote<br />

which ut ueed to evaluate ability or knowledge of -human befngm. All<br />

too often, hovever, there concepte are over-eanphaaized at the axpenee<br />

of a tart characterietic which ir coneiderably more fmportmt.<br />

A teet mey be perftc,tly reliable, yet mi&w~be measuring entirely<br />

nonvalid varianct, i,,e., it could bt measuring the wrong thing but<br />

doing it in e beautifully perftct vey. Content validity, very important<br />

for obvioue reaaose, ie occaei~nelly viewed by layum a8 the only kind<br />

of toet velidity that ie rtquirtd in Order to 18bel � ttet ne being 8<br />

good ant. Now the abrurdity of this point of view become8 mert apparent<br />

when one tnalyxtr the types of dtciefone that era going to be rnadt ebout<br />

tnlitttd men on the beet8 of their tart tcortt. Hart of the dtcieione<br />

are bared upon.an important aeeumption. The aeeumptioo made is that<br />

those individual8 vho acore high on the teat are performing better on<br />

the job (end thue should be rewarded), and those rcoring losrtr on tht<br />

teat art accordingly poortur perfonnere on the job (and hence ehould not<br />

bt rewarded). Whether th:Lo ir a true asrumptLon should not be left to<br />

cr:etal-bell garing or “ermchair” peychology, but � hould be subjected to<br />

verification. Thir vtrificetion mey be referred to as eetablishing the<br />

dtgrte of azpirical validity of a test, 1.6.. hw do ttst score8 actually<br />

related to en external mtesurt of on the .job pcrfonnenca? Ut et the<br />

Depertment of the Army ham rectrtly ‘incorporated e plen of v8lldetion<br />

for both the eveluetion taete ana the C-der’e Eviluetion <strong>Report</strong>.<br />

Detail8 of theta validation procedurea will be diecursed in mother paper.<br />

The purpose of thie paper io to describe � sptcific facet of the etudy<br />

which we cell tht “preliminary validity report.”<br />

The prtlimfnary velidity report ir Intended primsrily for the u8e<br />

of tt8t rpecialisto in the Test kvelopmunt Branch of the US Arm9<br />

tnlieted Evrluation Center. The purpore of the rtport is to provide<br />

guidelines in ttrt rtvieion bared on pertinent � tatfetic~l date which ie<br />

not aV8ilAble from our reguler ���� � rulyeie etetirticr. Thfr prtrtnt<br />

validation policy reprerente an improvement ovtr prevloue velidation<br />

offortr in that validity data covering the entire ttrt outline and<br />

indivfdurl test iteuu can be provided prior to teat rtvieion to allow<br />

fat their uee in the decirioa mrklng process of tort dtvelopocnt.<br />

,<br />

. .<br />

, - _-. _.<br />

.<br />

, ” *<br />

50<br />

---,- - . .<br />

._--_-_.


Teat reliability? -- Certainly 6OmC reliability 4-6 BlCCC66~~ in our<br />

test6 in order to obtain empirical --alidlty. Content validity -- again<br />

we era certainly concerned ,that itema in our tests should adequately cover<br />

the relevant arpect# of job behavior. Eut UC also feel thet there are<br />

only the minimum dcrsirable~ct~aracteri6tic6 thnt a test should have. Our<br />

earlier validation rjtudics asked the question: Doe6 this tC6t pO66e66<br />

anpirical validity? In the present validation procedure, thara 16 no<br />

lOtl@W M empharia !BOlely upon the qu56tiOIl -- he6 this tC6t pO66e68<br />

empirical validity7 -- but more directly urd more appropriatoly, the<br />

qUC6tiOKI is Why is the test valid? and Eov can it6 validity be improved?<br />

. .<br />

As noted in. th‘ra &k&ue paper, 'this one al;o reprer~tr ‘thi<br />

&pplication of itan selection technique6 advocated by CullikrCn in hir<br />

“Theory of Mental TlBSt6." Again, these techniques raqufrs the use of<br />

the point-birerial correlation coefficient for it--tC6t and ittsncriterion<br />

correlat Ions. The specific question v6 ark is, How do each<br />

of there itanr fnflrreoce the validity of the total evaluation test?<br />

The data provided for ita in the preliminary validity report of<br />

our validation sample inclu.de: (a) item p-values, (b) standard deviations,<br />

(c) itar vaclancco, (d) point-biserial itan-total test correlations,<br />

(e) point-birerial itan-criterion correlations, (f) item relitbility<br />

indexes, and (8) itcun validity indexas.<br />

The indexer of rcliability'ami'vrlidfty are obtained by multiplying<br />

the standard devfat:lm of each item by it6 rarpactiva point-birerial<br />

itenwtC8t correlation or point b18crial item-criterion corralatfon.<br />

They reprccrent the contributfon that the ftan maker to the reliability<br />

end the validity of the total toot, respectively.<br />

It is the last concept, the index of vrlidity that nw pO66e66e6<br />

special rignificanccs to our tc6t 8pcciall8t6. By inspection of the<br />

indexer of validity for each it-. they can detcnaine which itaru are at<br />

hart raeking a pori~cive contributim to the validity of the total tt8t,<br />

and inrure their inclu6fou in rubre~.uent te8t rcvirionr. If item<br />

validity indexer are available for each item that 18 includad in a<br />

test, it will be poerible to cetimate the total tart validity before<br />

it 16 � dmini8tered. The paper On itan Mdy8i8 hd6 irlicatsd hw telt<br />

-. .<br />

characteristics ruch a6 the mean, rtandard deviation, and reliability<br />

can be more closely ccntrolled. USC of the item validity index enable8<br />

US to exercire additional control over the mo8t salient feature of a<br />

tart, it8 ability to dircriminatc bctvrcn those lndividual6 vho 8re<br />

.<br />

good and poor performer8 on the job.<br />

51<br />

.


.<br />

.<br />

.<br />

.<br />

. -+_<br />

.<br />

. .<br />

*, .<br />

_ _ _ _ , ^<br />

.<br />

.<br />

.__ _ _.,_ _C..I---I.- -..-: .-.-- - .-._ ^ ._<br />

To fllurtratr how the indexor of validity for aach itaa lcflucoca<br />

the validity of the tote1 tart, consider tho follcwin~~ formlo,<br />

Ubrt4:<br />

ber thb validity of the<br />

‘xv tart,<br />

Zri,ot, the itma validity lndasi;,<br />

io tha point-bioerlal itemcriterion<br />

currolatioo<br />

multiplied by tho itrar<br />

rtmdrrd deviation,<br />

Bitei, ths itaP roltebility<br />

index, io the pointbioerial<br />

ItCE-tQot<br />

correlation luultipliod<br />

by ths itom otmdard<br />

doviotfon.<br />

Prom one cf our validity � ������ #the volfdfry eoefflcieot for tha<br />

tot81 test was .52. It wao found that the Cricoi two equal to 5.541,<br />

and Eritoi wab equal to 10.5’77. Entering the above equotlon YO find:<br />

‘w -<br />

5.551<br />

10.577<br />

Solving for the above, the rtV oqualo .525 which corraopondo to the .52<br />

which woo the volldity of th’o totol evaluation teat cmputod ftm the<br />

Powron product-nt formula. uoe of the indu of validity ptanotoo<br />

the � oundneoo of the “item-b,ank” concept,vhstetn numtouD prwfourly<br />

tried Itmoo pro readily wcsooible to the toot conotructor.<br />

Another urrful placr of information that vi11 bo provided thr teat<br />

� pmialloto lo a grophlc plot of the itaa-validity indexer and the itwe<br />

reliability indoxeo for each of the icmaa. One CM, thrroforo, tall at<br />

a glance vhich itano are contributing moot to teat validfty and taut<br />

reliability. Thio lo eooentlally tho oome kind of graphic dzrtration<br />

ahown by cu~likmn $n hio chapter on itm � olyoio.<br />

Row there ore the major changer in our volidotion � tudioo which vo<br />

bavr preomtly incorporated. Since Incorporating thio nmthcd, we have<br />

cuoplated or aeorly completed studier on � pproximtely oevon different<br />

mllitmy occupational opoci*litioo. Portunotoly, nearly all of rho teoto<br />

evaluated no far poorsrr � oam drgrar of aapirical validity.<br />

-- ___.-_--- _--~_ . .., __ .----.-...----- --. -___ __- . --<br />

.<br />

.<br />

52<br />

._ __ -.<br />

.<br />

;<br />

.<br />

c<br />

.<br />

.<br />

.<br />

.<br />

/<br />

:’<br />

\..<br />

’ :<br />

.<br />


.<br />

.<br />

One of the p~oblmnr that we probably all encounter in OUP taste me<br />

hiuh intrPcorrel4tionr baltween rubfact-matter 4~448. mture t o r t v~lidation<br />

affortr at USAF&Z will eneapparr further changes aimad of helping<br />

LO 4llOvfate thfr 8ittUtiOtl. !?88Qntirlly, -14 Are now plannin~~ a factor<br />

aulytic approach to telt. devrlopumnt. l-ho uhury4fin8P KwltlfoctoPl4l<br />

itam-snolytlc method has been ool6cted for thir purpoor. It 16 a method<br />

for sotfnrting f4ctor lordiago vithout the &rtai&bla tark of eauputing<br />

4 10+tx-125 v4ri4bls fnt,oPCOPP4l4tfOU WLtriX. A6 coon a8 ccxeputer prograplo<br />

4ro devolopod 4nd caarputor tio+o bocarnrr 4v4il4bl0, ths Evalu4tiou<br />

Section will rprt6mtic4l.ly bogin utr4cting group8 of itan that rhatld<br />

br rcorod tog&hat. We hlopa to obtain 8ovo~4l orthogonal rubto8t8.<br />

Tbfr approach h48 two m4for adv8ntagor for the Army. Pir8t, it<br />

8hould rerult i n an improwed diagnortic profila bec4ure mco~o8 uould<br />

tend to be independent of uch other end thu8 vould be awra meaningful<br />

and more e48ily interpreted. A profile tend8 to be of llrafted 000 if<br />

the vufour profile � ��� being 8COTOd are highly intercorrelated.<br />

Moxinnm utility o f profile .8cor48 demndr r414tiVOly unique varf4blo8,<br />

which m4~nr th4t e4ch variable OP to8t ue4 me4ru~4o 4 factor uaivac4lly.<br />

We are hoping that the Wherry-Wtnor approach till help u8 achiovo thi.8<br />

end. A second advantage fr thrt we till lncr6686 our knoulodgc about<br />

valid factora. If our Vo.riOU8 rubtest 4rQ uaivoc4~, the; masruram4at<br />

of tha moat valid foctorr cm be BlltphaOfhad which w6 hope will lead to<br />

wan gr6ator incPue68 in the validity of the tot41 ovalurticra te8t.<br />

In 8u11ana~y, the ~80 of tha prerontly incorpor4tod proliminrry<br />

validity x-sport and, ln puticul4P thr index of validity rhould help<br />

to inprove the urafulnrrr of our sviluatfon tolta and incPea80 the<br />

4ccurocy of porronnol docfrionr based upon thorn. Out long-rurge raoc&rch<br />

probleo till ba that of dotsnalning the porribility of obtaining orthogonal<br />

rUbtOrt vithln our 4VAlMtiOn te8t8.<br />

53 i<br />

1<br />

i<br />

f’ .<br />

1 1<br />

I<br />

---.--- .I.<br />

i<br />

i<br />

. /.‘..: i,<br />

(2 2‘<br />

.<br />

I


I<br />

.<br />

Guilford, J. P. Reliability and validity of ma&mre~. Facto? At;slysfs.<br />

~uu York: ~~Cr~-H~111, 1954, 470-538.<br />

Chilford, J. P. Validity of Mururewntr. Pundmmeai ctatirticr In<br />

paycholoay end eductrtion. Rev York; McGraw-Hill, 1956, 461-486.<br />

Gullikaen, A. Item eolectfon to cusfmixe teat vel:dity. Proceediaer<br />

of the 1948 invitet!loml confemnce 011 terting problme, Prinastoa,<br />

I. J.: Bducatlouai Tooting Sc~vIca, 1948, U-17.<br />

Wherry, R. J., md Wider9 B. J. A rcthod for frctotiog large nmbara af<br />

’ ittam. PspchaaatrilSf, 1953. 18, 161-139.<br />

-<br />

.<br />

54<br />

.<br />

.<br />

*<br />

.<br />

-._. - .- - - -__-<br />

.<br />

.<br />

. ,<br />

” i . i<br />

.<br />

.<br />

.<br />

.<br />

.


.<br />

.<br />

.<br />

The Prsdictioa of fob Proficiency<br />

ALEXANDER A,iLCNCO<br />

Staff Research Department of the Chief of<br />

Naval Air Training, Naval Air Station<br />

Menphia , . Tefmaoeee<br />

TM development of thia papar proceeded Qlong the line of b-craic to<br />

Qpplied rQaearch in the prcdfctlon of job proficiency. TtxImdlkQ, in<br />

Veraonml Selection” (1949). diatinguirhad between proficiency on the<br />

job per aa, and proficiency in job training. Only the latter type of<br />

proficiency ia conafdercd hero. Tvo araQa of roaQarch Qro cxamfocd:<br />

i.e., (1) b~alc re6eQrch on %otivQtion“ in e training emriroment; (2)<br />

applied raacarch on the predfctlon of school achlovemant. The firat QreQ<br />

involvca cozrelatiorml tcchniquaa which arc conaidrred to he Qn iatprwcd<br />

� ppro8ch to the araaurcmsnt of oonrptltude varfance. Included i n nonaptltudc<br />

variance, of murae, la a certain degree of Ymtfvatioa” variance<br />

vhfch haa been Qluaiva gentrally to amasuremant in a pure form. Howaver,<br />

aa we ahall ace, oum intereating thing8 can be dons in “wtfvation”<br />

maaauremant by the we of part Qnd partial correlation proccduraa. The<br />

� pplicQtlon of fnfonnation obtained from partial corralatlon procedure8<br />

to cndivfduafa by the usa of ratio derived scores wLl1 also be discuascd.<br />

l’tm aocond area involves the uae of multiple correlation and p866/fotl<br />

frequency procedutaa to ottain the beat prediction of technical training<br />

perfonaance. In thie regard, variour aclectlon md training grndea were<br />

866e8aad for their potentfal to 3 edict final trainfng performance.<br />

Operationally, thQ6e tvo approaches to the prediction of fob proficiency<br />

may be conridered to overlap, however, as vQ rhall 6Qe, that0 arc certain<br />

diatinctionr betwaca them vhich justify their indtpandtnt conaLdtration<br />

to the m6aaurmccnt of job p,rofLcfency. Several studier conducted by<br />

the Stuff Rercarch Dapartmmt Qt the Nav81 Station, Mnphir, on thcro two<br />

areaa will form the aubjcct matter of thfr paper.<br />

1 would llka to outlima the content6 of tha paper at this time in<br />

order to give you the focal pofnta of the mtcrlal I will be prorenting.<br />

I<br />

fie Hcaauramant and prediction of Hotivation<br />

A. Residual Cain<br />

B. Ratfo Xndex<br />

C. Motivation NQmsuramcnt<br />

D. Application of the Achievaaont Index (An applied<br />

rtasarch study)<br />

If The HtaaurQmQnt Qnd prediction of School Achisvannnt<br />

A. C,orrtlatioMl prediction<br />

B. Paar/Pail Odds to Prediction<br />

5s<br />

.<br />

,<br />

.<br />

. !<br />

i .<br />

I<br />

_<br />

.<br />

.<br />

1:<br />

. I


-IThe - Mearurc?ront and Prediction of Motivation<br />

By way of introduction to the rtudiar on the marure&zmt of moth-<br />

. ’ vatfon tvo 8tati8tical concepc~ nerd to be defined co fit our context:<br />

residual gain and actual to predicted ratio.<br />

.<br />

A. Reridual Cain. “Residual gain@’ ~88 dercribtd b y DUBOIS .in hir<br />

text on ‘Hultiv8riate ~relatjonal Analysir” (1957) a~ tht rcrldual<br />

vhich rtmtin, vharc the variance, vhich an lnLtia1 8core ha8 in CQmnOn<br />

vi th the flntl score, lr partialed out of tha find 8cort. Actually,<br />

Tborndibt, Brtgmrn, Tilton, and modyard in a text on “Adult Laarning”<br />

(1928) are credited for the firrt research employing reridual gain.<br />

Hirtorically, “re8ldual gaie~” va8 developed and used in a context of<br />

gain in learning. The clarric que8tion in laarning rertrrch has been:<br />

What are the correlate8 of learning?” This a88ume8 of cour88, that<br />

learning LJ not to be confurtd vith intelligence, � chievtmmt and other<br />

factors for which ttrtr have been devfrtd. mutver, several traditional<br />

probltma have been arrociatad vith finding the correlate8 of learning:<br />

(a) Aov can g8in be txpretmd to avoid metric difference8 betvren initial<br />

and final mearur4mtntr? (b) How can vt hold cormtent the known differtnct8<br />

among Individual8 in learning (or gain in achievtmsnt) in the fiM1<br />

msnrurement? a,& (c) Art the usual lw corrtlationr among gain eteaJUrbJ<br />

due to the multiplex nature oE gain or perhapr due to inadequate maJureJ<br />

of gain? “Reridual gain ,” derived by part correlatton techniquer, appearo<br />

to provide Jn Jdequate solution to each of there problem. hi8 may not<br />

sppesr relevant to the mstter in qwrtion here--nmely, tha mesourmmt<br />

of motivation in a training tnvironment. Houwar, the evoiution of<br />

“reefdual gain” AJ an index of learning ability va8 df*.tctsd eventually<br />

toward the w88ureumnt of cwtivation by a minor change in the baric<br />

dtrign utilized in the correlation81 rtudltr on learning. Thorofore,<br />

in order to trace the role of 8’remidual gain” in rsrearch on uativation,<br />

it fr necerrary to digrarr a little further on the nature of<br />

tBresidurl gain” itself.<br />

The baric paradigm utillittd in rtuditr on ~‘rteidual gain” fnvolv~~<br />

(1) pre-tertlng of achievement of rkill in a particular area. (2) Subrequent<br />

to thir, a given block of related inrtruction 10 introduced<br />

axparlrmnt8lAy, or ir provided by a technical rchool (~8 a logical rtquence<br />

to the pre-test itself, to the circuiwtancer dictate. (3) Lastly, a<br />

fl.nnl grade ir obtained for thir given block of instruction. The problam<br />

prerented here is to � ����� the degree of learning that took place<br />

behmen the pre-test and post-tart. lhir can ba done reveral different<br />

way8 (f.e., crude gain, X g,rin, etc.); however, “re8ldual gain” appurr<br />

to be a more oatlafactory rtatirtic of gain Ln 1earnLng.<br />

.<br />

. .<br />

. .<br />

.<br />

56<br />

.<br />

.<br />

.


.<br />

The basic fomula uhic:!~ generates the part correlation between<br />

“reoidwl gain” and a third variable is as follow:<br />

r(2.1)3 -<br />

‘23 - r12r13<br />

Where : 1 - Pratcot<br />

2 - Poottest<br />

3 - 0xridr V6rZable<br />

The boric fotlsuia vhich generate6 the fntercmrelction 5etwcrn cry t v o<br />

rrtr of residual mores, i.e.. 22.1 end 64.3 16 � 6 follour:<br />

‘24 - ‘12=14<br />

r(2.1)(4.3).- -<br />

- ‘23=34 + ‘12’13’34<br />

- r2<br />

l 34<br />

Where : 1,3 are inftial scored on two<br />

different test6<br />

2,d are final scores on tvo<br />

different te6 t6.<br />

Thf6 bring6 W CO the point, vhfch I indicated earliet, regard-w the<br />

tolr of “re6idual .-.6in” in ?ixotLvrtlon research. The! formulas ju6t dlscwred<br />

prWid6 a technique to divide the measured v6ti6nca fn rkil]. or<br />

echiwetmmt into two pert6: a part which is completely predicted by<br />

msrruted skill or schievement obtained at BOWi CCrffet point and & part<br />

which ie unpredicted by the e6tliet w66ure. Due to the mtute of the<br />

v6ti6bleo 6nd the baeic p6radfgm outlined above, the logic of the ritwtion<br />

leads ub to posit that the unpredicted vati6nce of “rClidu61” in tepre-<br />

6entative of “g6fn” in lermlng. Houevet, if instead of uring an achlevcmerit<br />

or skill mSa6uru a6 out pte-test we u6ed en eptftude utersute, then<br />

the logic of the situ6tlon would 146d u6 to porit that the rcrlducl fs<br />

more rynonymous with %mtivatfon” thsn “gain” a6 described 6bwe. ThU6,<br />

nothing has ch6nged except the n6ture of the pre-te6t. The vrlue of thir<br />

residu61 e6 a predictor and criterion both in the measurement and prediction<br />

of wtivation rill be illU6trated below.<br />

1. Retfo Index. Iha second concept of importance in our treatment<br />

of motlv6tlon tsseatch 16 the ratio of actual to predicted achievement.<br />

This hss been reforrcd to popularly 66: wer6chfcvementf underachievement ;<br />

.achievement index; dF6crcpancy 6core; PAQ (preparatory achlevemnt quotient)<br />

at cl.<br />

57<br />

.


.<br />

-._<br />

. .<br />

. .<br />

- .<br />

.<br />

. *<br />

Duboie (1957) hor rlhovn thet, ‘*e properly conetructod retie (ouch<br />

u that described ebove) te eaecntlelly � opeciel cane of a reelduel,<br />

i.e., the velue in the rmrnoretor vorfeble lees thet pert wkfch fe cortolated<br />

vich the denanioator variable. Accordingly, a rotio � hould<br />

corralaza .OO with the dcnaminattir verieble.” Thin hoe practieel<br />

� ����������� fa that ue ten perfom our rcteaerch in tha form of port<br />

and pariiel corraletion technlquae, end utiilee the results, if lndlutad<br />

by the dete, la the form of a ratio Index. Traditionally, thfo<br />

index has been derived by dividing � atudent’e, obearked performance<br />

(i .a., any training grede) by hie predicted parformeaco. One could<br />

divide by � etudent’e rw eptftude grede, but it ie coaoiderad to be<br />

8 further reflnmeeat to luee e predicted score derived on the beeie of<br />

OQ@O � ptitudr grede. In � lthar ceee. the reeultent index 18 untorrelated<br />

tith � ptitude (i..., .OO). Thur. rlmller co the pert correletlcm, the<br />

retfo epproech dlvidec m trelning grede into tm pertr; thet verfrnce<br />

rrhich fe predicted by rpritude and that verience which is unpredkted<br />

b y aptitude: (i.e.,the achlavemeut index). 7714 ochiavemaant index<br />

will be 1.00 or < ) 1.00 dcpandfng on the megnltudee of the cmmrator<br />

end denmfnator � ������<br />

It me7 be benoflcfrl to clarify two uuttere at thie point in order<br />

ta define thlr index properly. Tha achisvaneat index 1s oiraller to the<br />

“reelduel goin” AS we Fndfcetsd. The former meeure ie indicative of<br />

"mot ivat ion ,I’ end the letter, of “gain” In leeruing. One rae7 Infer<br />

thet both maneuree Ore “pure” indices of motivotiou end gain in learning.<br />

his ehould not be � eeumed *inc.> they era both roeiduole repreeenting<br />

verieace unexplelned by mserured � ptitude on the one hand end athlsvement<br />

on the other. Dubole In “Correletlonel Analyrie i n Training Reeeerch”<br />

(APA Peper, 1956) and Hayo in “PAQ Kenual” (e CNAl7XHlXA Reeeerch <strong>Report</strong>,<br />

1957). rtepactlvaly, indfcete that each residue1 lo not � “pure” meeeura<br />

of Rein or motlvetloa.<br />

The � econd problem often encountered concerue the predlctLve<br />

cepeclty of the � chlevcmmt index. This cm ba viewed in tvo ueye. On<br />

the one heod, � trefnlag ;grede , � e I indiceted, emi be divided into tw<br />

pert8 which are uncorraletad vith � rch other. Therefore, in � correletlonel<br />

� enee, eech pert connrrlbutee uniquely to the prodlctlon of an<br />

outside criterion. Houevcsr, rfnce these two pert8 era breed oa tha aeae<br />

trrlning grade, tt � hould not be � rmmed thet the cwbiood prrdlction<br />

by there perte will be grtrrter then prediction by thet trelnLag grede<br />

iteelf. This vae diecurecd in � reeanS f3UTECHTR.A rorcerch report by<br />

-go titled “An ApprLerell of Retie Scores.” Date on thlr mtter ir<br />

conteiaed In Tcible 1. TV0 rchoolr era reprsrrnted. It ten be rem thee<br />

the � chievcsnt Index (elno celled PAQ) dose not add to the Multiple ‘1”<br />

AS coutpered to the rfmple “r” vtmn uring t h e training grade clone. Thue ,<br />

without PAQ the OLmple r’(i wre .761 end .767; vlth PAQ, the Multiple<br />

B’S vere .762 end .763, rcepcctfvely. However, there ere other prectieel<br />

_ ,.. -- ---- . . .-__ ^_. ._..- __ - -- .._-<br />

I<br />

.<br />

.<br />

.<br />

58<br />

.<br />

i<br />

‘.<br />

/- ;<br />

I *


.<br />

- -<br />

.<br />

.<br />

.<br />

. *<br />

.<br />

i<br />

’ f<br />

“. . z<br />

~ -<br />

. .<br />

. I-<br />

. . .<br />

.<br />

t ”<br />

. .<br />

.<br />

. .<br />

-, 1. _.- j .-._ _.. -.A-- -<br />

,<br />

.<br />

reamn8 for dividing a tr~fning grrdr into tvo partr. Tt~ir iavolver the<br />

a$e of the achievmmnt index for dlagnortic pwpooer end will be tremted<br />

in beetion D to follow. -<br />

.<br />

Teble 1<br />

A Comparfson of Simple and Hultipla Correlstion G~tnp<br />

the Ratio InGax for Am(A) end ATR(A) School8<br />

-Ma(A) ml(A) (414) .767<br />

PAQ-AFU(A) ATR(A) (414) .763 '<br />

,c. Motivrttioa Xoamrownt. Earing defined out working coneopts,<br />

of “reoldunl grin” and uchlrvmaent retie, I would !ika now to present<br />

eme rsrearch findlngr dnvolvi~g the me of both rpproechar in aotfvation<br />

reararch. The rerurch VB# conducted by the CXATRCRIIU Strf f Reaurch<br />

Depertmnt located at Hmphie.<br />

1. PAQ Wmocrl. Iclyo, C. D., ClUTgCRTRA Rereuch <strong>Report</strong> 1957.<br />

a. Ratare of AchiovrPont Index:<br />

me firrt quartioo roleter to the ruture of thr achlevemont<br />

index. A# wo fndlcrnted, the logic of the paredigm vhereln eptitude<br />

ir partialed out of a trelning grade luda UQ to hypothuine that the<br />

vuiance rmainfng ir related to motivotioa OT effort-put-forth. Of<br />

comae, �� � lro lo rocogaited that othor factora contribute to thlr<br />

unpredicted nrlmce: 1.0.~ error, parromlity dffferencer, lnterert<br />

end � ttltudee, and pteviour � chlovmeat la � porticulor orea aad ability<br />

not auusured, � tc . l%r tuk, tharefore, ir to detormino uhat axteat<br />

motivation ir contefned in the reridual vuianco.<br />

The method edoptod to roroloa this quertltm was tho u8e of paat<br />

r*tlnge. Tbo clerric UBO of peer rrtitgr vu employed uhereln clrrres<br />

of 15 to 25 uen wro erktrd to cuminate and rank thrso rtudaatr In their<br />

59<br />

.


. . - -._<br />

.<br />

.<br />

.<br />

. . a<br />

- -- ..-. - _. _.. .-_._ .1 ._-.- ---- --.--- - -- __._.. ..__ __.. -___. ___<br />

roctioa who vero trylag the hardert to meter the courno and, vice verea<br />

for three mn vlro vero trying rho lwrt hard. Thhs objective VW to uma<br />

aa early � chlweaeat iadut to predict effort en � �������� by peer rqtbgr<br />

later in training. (Ttm rooulrr are tontaiaod In Table 2u.) The carrelationa<br />

of them two mumura~ range fraP .17 to .40 for four tcchalcrl<br />

trdning ochoolr and are trll rlgnificant.<br />

trblr 2r<br />

Correlation of Achiovemmt Index vith<br />

Paor Rctiagr on Effort<br />

Avfat loll School w r<br />

1. h(littO biOCh&AiC 166 .40<br />

2. Structural kkchmic 161 .U<br />

3. glectrmico Technician 209 .17<br />

4. Training Dovicemn 266 .39<br />

.-<br />

Since peer r-.tlngr th.melvee ue not imum to haio effactr for<br />

rwaon~ � uch � # intelligence or clorr achfavment, further InvoatigrtLm<br />

wu made to remove the effect of rrtLngo on InteAlfgmao frar the<br />

correlotion of rating on rtffort wd the � chievumnt index. Ttlir was<br />

done on A ample of lb6 rtudmstr in the Engine ktachanic Schwl at ntnphie<br />

(cf. Toble 2b). ;he rowkant pertlal corroletion VU indicated to bo<br />

*lb. When compared to the original correlation of .40 for thoro #am<br />

rariablsr vo cun 8~ that halo did lnfluenco ratinge oa offort - but tho<br />

portlol corrolotion vu #till � ignlficont of tho .OS 1~01. Uhilo mro<br />

roroorch nudr to be conducted In thle � �� tt can bo raid that ea~ of<br />

the vsrianco In the achievement index ha8 been idmtiflod 01 bring<br />

related to offort.<br />

.<br />

60<br />

.-..___ ._.- _. __ ._ -.. _<br />

. *.<br />

.<br />

e<br />

.<br />

. .<br />

.<br />

.<br />

,


--<br />

1<br />

. *<br />

__-_..___ __.-_.- I .<br />

� �<br />

. .<br />

�<br />

._ _, _.-_l ____- -..c--..- -_.--- . ..-- .___ll-l.<br />

Table 2b<br />

Intrrcorreletkms kaong Pear Rating8 on Effort end<br />

lntelligmce end Achievmmt Index<br />

1. Peer Rrtlngg--ori Xntelligenee<br />

2. Peer htingr 00 I!ffort<br />

3. intelligence (#my CCT)<br />

4. Aehievaent Index<br />

b. Stebilicy of the Achievasaat Index:<br />

1. 2. 3. 4.<br />

- .66 .46 .46<br />

e .29 A0<br />

m .lO<br />

The � ecxmd quution invclvea the relfrbility of the<br />

ehievement index. fafcrrmti~ on thle quartloo. ~08 drawn from enother<br />

� tudy Involving 196 8tUdMt8 for the Aviation Structural Mechanic School<br />

at Kemphis. Correlations wre obtofnod brtww tha achievePant index<br />

derived urly in training end at the end of treiaiq in two 8chooli.<br />

h e COlTOl~tiCm8 cbtelmd wro .69 and .35. fii8 indlcatu a noderate<br />

atability in the achfevtnant index. At leeat, it indicator thet the<br />

re8iduel fr not compared entirely of random error.<br />

2. Uotlvation Uea8uranmt. tteyo e n d Umnning. l!ducetiaul and<br />

P8ychological Hsrrurazmx. Volume 21, #usher 1, 1961.<br />

lhir rtudy invertl~.&tul enothor bark arpect of the � chieveeeat<br />

index, nrrnoly the prediction of � rubrequent � cklevaent index by e<br />

orriety of motivation meaoure# Including en early achlevment index.<br />

Actually, the echievanmt index in thlr inrtanco took cho form of a<br />

part corroletion fnrtud of e ratio of obrerved to prullcted achievemmt.<br />

A8 we indiceted before, there WO 8tetfetiC8 � re 0888ntldly<br />

rimiler.<br />

The mocLvrtfon meaeure8 uaployed wore:<br />

a. Peor retlng cm offott<br />

b. Self r*ttngr on effort<br />

C. Chock llrt o f rtudeat behavior<br />

d . Pietorte Nearurea<br />

-._ ----- ----I’<br />

.<br />

,<br />

61<br />

.<br />

, .


.<br />

.<br />

e. Grader in an urlfer couree wizh aptitude rauoved.<br />

(Thie is the reeidual which ie basically einilar<br />

to the retfo achievwsent iodbx and wa6 tentatively<br />

ueed a6 e mexeure of motiwtion.)<br />

The analpele proceeded in three etepe. Piret, fntercorrslotione<br />

meong the ten variablee were obtained (cf. Table 3). Secondly, the<br />

three aptitude meaeures vare partfxled out of each of the reaefning<br />

variable6 which wore thn intercorrelated u partial8 (cf. Table 4).<br />

Leetly, multiple corroletione vere canputed batween the raeldualfred<br />

motivation and echool grade predictor6 on the one hand rad the reeidualixed<br />

eriterlon variablea on the other hand (cf. Table 5). A8 you<br />

��� � ee, In tha lattar table, all the predictor raeiduale combine to<br />

predict both criterion roelduale very well. The maximnn multiple lo<br />

a70 for criterion #I, aud A4 for criterion i2, however, the prisnxry<br />

variablee involved in the multiple partial correlation were the peer<br />

rating residual and the fundemeotale echool grade reeiduele; i.e.,<br />

4.123 and 8.123, which yfeldcd~emltiple partial correlation8 cf .67<br />

and .41, reepectlvely.<br />

Table 3<br />

Intotcorrelatio=a Ffeme and Standard<br />

Deviation6 of Verfater - (H-196)<br />

Ueaeure 1 2 3 4 5 6 7 8 9 10 Me&ii S.D.<br />

::<br />

43:<br />

5.<br />

6.<br />

2<br />

9.<br />

10.<br />

cm 29 29 11 -09 04 15 36 33 31<br />

AR1 04 00 -10 03 08 38 20 -01<br />

HECH 16 08 10 05 22 43 38<br />

Peer Rat inge 28 19 00 30 59 34<br />

Self Rating6 -03 -10 09 19 04<br />

check Llet of 02 11 22 22<br />

Student Behav.<br />

Plctorlel Motlv 09 01 03<br />

mch Fund 57 40<br />

Sheet Hetel Unit 61<br />

Welding Unit<br />

tiote: Decixul point6 aeitted<br />

55.61<br />

55.54<br />

56.26<br />

29.24<br />

30.20<br />

97.71<br />

7.18<br />

81.70<br />

79.37<br />

82.34<br />

--- .._- _-.- -~<br />

. . .<br />

.<br />

. . .<br />

.<br />

.<br />

62<br />

*<br />

6.61<br />

5.82<br />

6.76<br />

9.60<br />

9.49<br />

11.47<br />

2.27<br />

6.44<br />

4.89<br />

5.24<br />

. .<br />

.<br />

.<br />

-y i


. . .*<br />

.<br />

- ..-.. ___ -<br />

--- -___ -.__- FI_---wmw--L. - --. -..h-.-.__ .<br />

.<br />

l4eArura<br />

Tab18 4<br />

fntArcorraltrtfonr of XotivAtLon and Criterion<br />

kftmmrei, Aptitude PertiAled Cut.<br />

.<br />

x4. 123 ‘5.123 ‘6.123 ‘7.123 ‘8.123 x9.123 ‘10.123<br />

X4.123 Parr BAtingr, -<br />

29 18 -02 29 60 30<br />

X<br />

5.123 Self RAtbIg -03 -09 16 18 05 !<br />

‘6.123 Checklirt 01 09 20 20<br />

‘7.123 Pictorial HUB. 03 -04 -02<br />

‘8.123 Uech Fund 49 35 '.<br />

%.123 Sheet !tOtAl Unit 43 i<br />

‘10.123 Welding Unit ”<br />

Note: DeeFePel points amfttsd<br />

Table 5<br />

Hultiple - PArtfAl CorrelAtioa Coefficlentr<br />

Betveen HOtiVAtion and CCitSrfA VAriAbleA<br />

MtiVAtioO #eArUreA Ctitsrla<br />

Sheet MetAl Unit Weldiog Unit<br />

x4.123, xS.123, x6.123, IK7.123, ‘8.123 .70<br />

‘9.123 ‘10.123<br />

. x4.123, x5.123, x6.123, ‘K7.123 .62 .34 i i<br />

.<br />

‘4.123, ‘8.i23<br />

63<br />

.<br />

. -. / .7<br />

./ .i :<br />

,>:<br />

c<br />

LX,<br />

;<br />

i .<br />

i<br />

i<br />

i<br />

I


,<br />

- .- - . . ---- .-..<br />

--.<br />

-<br />

D. Application of tha Achievamant Indax. I would like to turn now<br />

to 8 study on an application of the � chievanont ladax in a training situation.<br />

A8 we Indicated, the � chievmnant index can most easily ba darlvad for<br />

individuala by men8 �� � ratio of. actual divided by predicted achievomeat.<br />

1. A Note on Under sod Ovarachiavcmunt. F’rochlich, A. P. and<br />

ttayo, C. D. Par8unr;el end Gu:ldanca Journal, Merch 1963.<br />

This ~jzxnel rrticle rcp:rorantr a susxmary of the raraarch on the’<br />

� chiovammt Index and ohoutd be � ppropos al80 to the termination of my<br />

own remarks on thin subject. Froahllch indicates two applicctionr of<br />

the � chiavamant index (a) Bareuch md (b) Prediction. We have dwelt<br />

at langth on the potrntlal thlr Index haa for purpoaar of baric rsrearch<br />

on learning gala and motiuatImon depandtng on the nature of tha variable(s)<br />

partialad out; i.a., whathar ,achiavmant or aptitude 18 ra8lduallrad. As<br />

regards It8 predictive aspacts, we elro heva indicated that the achlevaindex<br />

doe8 pradlct training criteria but doar not add to the training<br />

grade. Ae notad, it repromate the unpredicted variance of a training<br />

grade which, whan corabinad with the predicted variance of that grade<br />

sat8 the multiple I%” a8 quel to the rtmple “r” (by urlng the trslning<br />

grade alone) In predicting en external criterion. Proehlich cleerly<br />

point8 that fact out, but emphasized that the marit of this dichotomy<br />

liar in It8 iroletlon of variance vhich wa8 not predicted by the ability<br />

maa8urtr utilized. Thu8, .when derived aarly in training, tho � chlevantmt<br />

indax 18 conridarad to hava marit In counseling “low-achiaver8” (i.e.,<br />

those who scored 1.00). Iha study indicated that, “if tha unpredicted<br />

pert of academic achiewment measurer tnotlvationel varience, it might be<br />

bvpotherieed thet if underachla~var8 of various ability levelc could be<br />

motivated to put out more effort,<br />

heve more interest, or t8ke on more<br />

positive attitude8 toward their studfe8, Sane might im;*‘ovc their final<br />

gradas .‘I<br />

In resume, it can be reid thet the achiavemant Index hao 8 graat<br />

potential for basic research in tha difficult area mearurlng non-aptitude<br />

variance ruch a8 motivetion. The pertiel end part corrclatlonr offer a<br />

good technique to isolate andi identify the corralatom of this rariduel.<br />

SOW use of this Indax in tha form of a ratio ha8 been made in uavol air<br />

technical tralnfng for counsalinng purpor*r. It is conrldarad that further<br />

research Into the netura of the � chlavrsnent index will prove to<br />

extend It8 role in an applied sannre.<br />

II Prediction of School Achievamant<br />

The recond approach to job proficiency maarurar i8 the pradiction<br />

of school achlevamant. This anployed the cl~s8lcrl regrarrion<br />

end pens/fail frequency proctcdurar to derive prediction8 and sppronimate<br />

odd8 to succeed various ochools in the Naval Air <strong>Technical</strong> Training<br />

Comand. The vrriebler utll:itad arc the � veilebla relaction end treinlng<br />

grede8. The technical School8 involved were the machaolcal and avionic8<br />

ratings. Briefly, I would like to prarrst an axample of the correlational<br />

teblt8 end al80 the parr/fail frequency teblas and their role in tha<br />

quality control of traiatar.<br />

.<br />

, .<br />

64<br />

__ ._ _ _ . _ .._ .---- - ___. -__ _--<br />

.<br />

. .<br />

‘.<br />

‘<br />

___-- -.--I--.---.<br />

e .<br />

.<br />

*.<br />

a<br />

i<br />

a<br />

.<br />

/


_ . . . _ .- ._ ..-.-_ _---.---- - -_____ --- ._ _- -.. ..-<br />

. .<br />

.<br />

.<br />

---- -.-.. __-.-.- _.___.__. . ..- . ..--_.-__-- .._ _ -. _. ._ . ___._ ..-L-.-L ._ .<br />

A. Correlational Prediction:<br />

T8bler 6 8nd 7 cent&in the correlatlonr of the trrining grade8<br />

utillred at two different point8 in the mme school (IW-A). These<br />

correlatlonr yielded the regrerrion equation8 nccersorg to dorive the<br />

predicted 8core8 contained in figurer one and tuo. Prediction tableo<br />

ruch � r.there, rather th8.n regreorion equ8tioa8, greatly arolot the<br />

training officers to gra.rp the mewing and uoo of predicted scorer.<br />

It8 application, of cour88, con8f8tr in: 1. Vetezmiakg the a:udent’r<br />

grades on the vari8bler acro88 the top and left aide of ffgurcr one<br />

and tvo; end 2. Locating the predicted school final average within<br />

the prediction tabler.<br />

.,*<br />

Table 6<br />

fntercorrclatlons of Variables Utilized<br />

, at the End of Aviation Pamiliariratloa School<br />

m(A)<br />

(N=1342)<br />

2 ,’ 3 H u<br />

BTB 1. .374 .392 177.63 12.32<br />

APAn 2. .478 76.72 7.03<br />

AFu(A) 3. 73.80 9.73<br />

Table 7<br />

Intercorrelation of Variable8 Utilized<br />

at the End of Phase 1 of APU(A) School<br />

APu(A;<br />

(N-1322)<br />

2 3 4 H u<br />

BTB 1. .365 .367 .381 177.79 12.29<br />

APAn 2. .389 .461 76.86 6.94’<br />

Phaee 1 3. ,851 76.27 9.01<br />

Am(A) 4. 74.09 9.32<br />

.<br />

.<br />

65<br />

.<br />

,<br />

,I<br />

- .^.._ -..<br />

, ;<br />

I


.<br />

.<br />

. .<br />

. .<br />

.<br />

.<br />

--. _ ._. ._ . ^_ . - --_._--~---.“~-- --..- .- - - _._ - _. .-_. ___ _.-_<br />

200<br />

195<br />

190<br />

185<br />

180<br />

175<br />

* 170<br />

c-4<br />

' 165<br />

m<br />

2 160<br />

ok<br />

2 155<br />

2 ;;:<br />

150<br />

$ 145<br />

140<br />

135<br />

130<br />

125<br />

120<br />

R - .530<br />

~1.23-8.23<br />

Figure 1<br />

AFti Prediction Table<br />

(?P1342)<br />

AFAU Ffnal Average (X2)<br />

-. --<br />

53 56 59 62 65 68 71 74 77 80 83 86 89 92 95 98<br />

66 67 69 70 '72 74 75 77 78 80 8? 83 85 86 88 89<br />

65 66 68 69 71 73 74 76 77 79 81 82 84 85 87 88<br />

~....<br />

64 i65 67 68 ';O 72 73 75 76 78 80 81 83 84 86 88<br />

.--a4<br />

63 64 j66 67 69 71 72 74 75 77 79 80 82 83 85 87<br />

62 63 i65 66 (58 70 71 73 74 76 78 79 81 82 84 86<br />

m-o-(<br />

61 62 64 I65 (57 69 76 72 73 75 77 78 80 81 b3 85<br />

"'Y<br />

60 61 63 64 ;66 68 69 71 72 74 76 77 79 80 82 84<br />

8<br />

59 60 62 63 i65 67 68 70 71 73 75 76 78 79 81 83<br />

----<br />

58 59 61 63 64 I66 67 69 70 72 74 75 77 78 80 82<br />

t<br />

57 58 60 62 63 i65 66 68 69 71 73 74 76 77 79 81<br />

-.-.*<br />

56 57 59 61 62 64 i65 67 69 70 72 73 75 76 78 80<br />

-0-q<br />

55 56 58 60 61 63 64:66 68 69 71 72 74 76 77 79<br />

:<br />

54 55 57 59 60 62 63:65 67 68 70 71 73 75 76 78<br />

!-,<br />

53 54 56 58 159 61 62 64:66 67 69 70 72 74 75 77<br />

:<br />

52 53 55 57 ,58 60 61 63 i-6: 66 68 69 71 73 74 76<br />

51 52 54 56 .57 59 60 62 64165 67 68 70 72 73 75<br />

-0.<br />

50 51 53 55 .56 58 59 61 63 64: 66 67 69 71 72 74<br />

--- - .<br />

.<br />

. .<br />

Regrerrion Equation: ~,,.196Xl+.53u[2-1.77<br />

Outtfng Score: 65 (i.e., oddr of 2 to 1 to pare)<br />

Xl-Basic Teat Battery Composite<br />

X2-A.PAIf Final Average<br />

E3=AFU(A) Pine1 Average (predicted)<br />

.<br />

,<br />

66<br />

*.<br />

c<br />

.<br />

.<br />

.<br />

.


__ ._ __ ._ .^ . .._--- -. ____. ..____ ~-__----- ---I -_-- .__.-_ - .___-_ _-.-. - . -.<br />

. *<br />

. .<br />

-----<br />

::<br />

a’<br />

_ ; ____ __-_.___ _ _s .,--1 .. ..-Clr-<br />

_--- --..-- ._.-.__C .____<br />

_ --‘-..--<br />

98 58 61 63; 66 69 71 74 76 79 81 84 86 89 91 94 96<br />

95 58 60 63; 65 68 70 73 75 78 00 83 85 88 90 93 96<br />

92 157 60 62; 65 67 70 72 75 77 80 82 85 87 90 92 95<br />

89<br />

86<br />

83<br />

80<br />

%<br />

a 77<br />

i 74<br />

2 71<br />

lj 68<br />

i 65<br />

62<br />

59<br />

56<br />

53<br />

50<br />

Figure 2<br />

MU(A) Prediction Table<br />

(R-1322)<br />

Fbue 1 Fin81 Average (Xl) i<br />

53 56 59 62: 65 68 71 74 77 80 83 86 89 92 95 98<br />

57 59 62i 64 67 69 72 74 77 79 82 84 87 89 92 94<br />

(1-m*<br />

56 58 61 63;66 68 71 74 76 79 81 84 86 89 91 94<br />

:<br />

55 58 60 63;65 68 70 73 75 78 80 83 85 88 90 93<br />

:<br />

55 57 60 62; 6S 67 69 72 75 77 80 82 85 87 90 92<br />

:<br />

54 57 59 62j 64 67 69 72 74 77 79 82 84 87 89 92<br />

----I<br />

53 56 58 62 63 ;66 68 71 74 76 79 81 84 86 89 91<br />

:<br />

53 55 58 60 63 ;65 68 70 73 75, 78 80 83 85 88 90<br />

:<br />

52 55 57 60 62 :65 67 70 72 75 77 80 82 85 87 90<br />

:<br />

52 54 57 59 62 :64 67 69 72 74 77 79 82 84 87 89<br />

c ---I<br />

51 53 56 58 61 63;66 68 71 74 76 79 81 84 86 89<br />

:<br />

50 53 55 58 60 63:65 68 70 73 75 78 80 83 85 88<br />

:<br />

50 52 55 57 60 62165 67 70 72 75 77 80 82 85 87<br />

49 52 54 57 59 62i 64 67 69 72 74 77 79 82 84 87<br />

---a,<br />

48 51 53 56 58 61 63 i 66 68 71 74 76 79 81 84 86<br />

R - .863 Rcgrsrrioa Equation: ~3'.8388x1+.210x2-6.55<br />

01.234.82 Cutting Score: 64 (i.e.,oddr of 2 to 1 to paw)<br />

Xl-Fbue 1 Final Average<br />

~-pAFAH Final Average<br />

X3=AFCl(A) Final Average (prsdlcted)<br />

.<br />

.<br />

67<br />

!


--- .._ __. __^ _ __-.-_ -.- -- _<br />

.<br />

- -- . . . .- :.- -. -. . _... .- _ _ __ I - ._. -.<br />

.<br />

�<br />

��<br />

� � �<br />

.<br />

.<br />

. . .<br />

m<br />

- .---- . .I . ..r _ -. _.-_-1---- _ -- ._ _<br />

B. Parr/Pail Oddo of Predfctloa:<br />

�<br />

no second mrthod ured to deterarlee a 8tudent’s chaxice8 of<br />

succeeding in training is bared on the observed Parr/fail frequanctar<br />

for vulou8 intervals of grader moat correlated with trainlog parformancr.<br />

Tables 0 and 9 ot tha handout contain the parr/fail frsquancy<br />

data for 2 variable8 taken at differant point8 fn tr&fntng fn<br />

the Avionic8 Fuadammtalr School et Hsmphlr. Data In tablo 8 io based<br />

on an “N” of 1340 rind a correlation of ,481 and data In table 9 on<br />

8n “W* of 1291 and 8 correlation of 25. The pars/fail frequsncleo<br />

are trmrlatd into approxiaate odde to succeed that particular 8chool.<br />

Thr ame data MI translated into two graph8 to facilitate the tratnlng<br />

officer8 interpretation and application of thaia pars/fail frsquancy<br />

tables. There graph are contained In figure8 3 and 4 and are relfarplanatory,<br />

tith the acmplor given vithin ach table.<br />

The predicted stoma, based on selection and training grader hmvs<br />

proved to be a valuable index to ultimte training porfomunce. The<br />

application of there predictions is the


�<br />

Teble 8<br />

Thr Probubilitier of Grrduation MU(A) School<br />

after a Student Carcpleteo APAX School<br />

(H-1340)<br />

AFM x A&) WJ(A) Apprax’. Oddr x -<br />

Pfa.Aver. of Total Graduater Attritqr for 8g8Fnrt Graduating<br />

9 4<br />

90-93<br />

86-89<br />

82-85<br />

78-81<br />

34-77<br />

70-73<br />

66-69<br />

62-65<br />

58-61<br />

54-57<br />

-53<br />

.4<br />

2.4<br />

7.6<br />

16.3<br />

20.0<br />

21.0<br />

17.4<br />

9.7<br />

3.4<br />

1.2<br />

.7<br />

. l<br />

352<br />

97<br />

213<br />

254<br />

244<br />

178<br />

95<br />

31<br />

:<br />

5<br />

5<br />

14<br />

5375<br />

35<br />

1’:<br />

5<br />

1<br />

5 0<br />

0<br />

2i 1<br />

43 2<br />

18 1<br />

13 2<br />

1<br />

z 2<br />

2<br />

i<br />

z 5<br />

-7<br />

0 1<br />

100<br />

100<br />

95<br />

98<br />

95<br />

87<br />

76<br />

#73<br />

69<br />

ii<br />

00 -<br />

llS8 182- 6 1 86 - -<br />

Table 9<br />

l’ho Probebilit%e$ of GrAduatfng AKl(A) School<br />

����� � Ltudent Comphter Phare 1<br />

. (N-1291)<br />

PhAre 1 % m(A)<br />

Pfa.Aver of Total Credoate<br />

94.<br />

90-93<br />

86-89<br />

82-85<br />

78-81<br />

74-77<br />

70973<br />

66-69<br />

62-65<br />

58-61<br />

W-57<br />

2:<br />

9.4<br />

13.9<br />

19.7<br />

15.9<br />

14.7<br />

9.0<br />

6.5<br />

3.6<br />

.9<br />

22<br />

61<br />

121<br />

179<br />

252<br />

202<br />

177<br />

91<br />

41<br />

11<br />

1158<br />

�<br />

�<br />

,<br />

69<br />

Attriter<br />

�<br />

- -<br />

Approx.Odda x<br />

for Qafn8t Cradu8ting<br />

22 0 100<br />

61 0 100<br />

121 0 108<br />

179 1 99<br />

126 99<br />

67 : 99<br />

14<br />

93<br />

18 :<br />

78<br />

bV@Il 49<br />

1<br />

24<br />

- 1 1: - - -8- 9 1 90<br />

I.<br />

i<br />

i<br />

; I<br />

i<br />

.<br />

i<br />

I(<br />

.<br />

; i-<br />

t :<br />

i<br />

� �<br />

�<br />

i


90<br />

80<br />

70<br />

60<br />

10<br />

.<br />

A,ctual da<br />

I<br />

I Exawpl~: 64<br />

2. Interrect probability CUNQ<br />

1<br />

I ,3. Interpret odd8 for •UC~BID<br />

I Grade 64-7 to 3 to pur<br />

I<br />

I<br />

00 7V 80 90 il<br />

AFAX ?irul Avor*go<br />

Ptgura 3<br />

8umaary utilf?y chart of probability date<br />

coatrlad in T&h 5<br />

70<br />

, _ _-_ - .._ .-.-_ .<br />

.<br />

Odds<br />

GVSn<br />

2 to 3<br />

3 to 7<br />

1 to 4<br />

1 to 9<br />

.<br />

.<br />

.


60<br />

40<br />

30<br />

20<br />

10<br />

1. Identify PtlAEr 1 grrda<br />

Exam$ls: 66<br />

2. faterroct probability curw L<br />

3. Interpret oddr for eucce88<br />

orode 66-3 to 2 to par<br />

I I<br />

70 80 90 loo<br />

Fhsra 1 Final Avrrrgr<br />

Piyre 4<br />

Gummy utility chart of probability data<br />

contained in Table 6<br />

. .<br />

. .<br />

.<br />

.<br />

71<br />

.<br />

.I<br />

Oddsl<br />

9 to 11<br />

4 to 11<br />

7 to 3<br />

3 to 2<br />

EV&l<br />

2 to 3<br />

3 to 7<br />

1 to L<br />

1 to 9<br />

.


.<br />

,<br />

.<br />

.<br />

rteferencer<br />

Dubois, P. R., Multiv8rlate correlational analylfe. Ww York Harper &<br />

Brothers, fmt .<br />

buboir, P. B. h t4aml~, u. H., Methods of res-h ln tcch~ical training.<br />

QIR Contr. # ~cear 816(02), NATT, ~av. Ed. ~e.ch. Rapt. 83. April 1961.<br />

lroehlich. H. P. and ?fayo, C. D., A note cm undar md overe&iewmat.<br />

perronnal and Cuicirnca JournalI Mrch 1963.<br />

L0DfJ0, A. A., etioo of rtudant performance Ln five wchurlcrl<br />

8ch001.. v rm<br />

Iennsrrce, December 1963.<br />

lougo, A. A., An appraisal of ratio meores. Bc,earcb <strong>Report</strong>. Staff,<br />

CNATECXl=iU, NhS wznphi8, feaa~aea, Much 1964.<br />

Lwgo, A. A., The pradactioa of rtudrnt parforranca in oaven avionica<br />

re: oolr . Berearch Rsport. Staff, m .w-i00<br />

Tameawe, August 1964.<br />

Uayo, C. D., PAQ Hawal. Renaarch <strong>Report</strong>. Staff, C2UfeCEmUr tU!I<br />

Nemphia, TenncBsee, 1957.<br />

)slyo, G. D. and Xauning, U. Ii., Uotivatlon meaBut-t, Educational,<br />

end P8ychologfcal Xeaaur-nt, XXI, 7343, 1961.<br />

‘Lhorndlka, E. L.; Brsgmaa, J.; tilts, J. U. and Woodyard, I!., Adult<br />

learuing. New York: UacMlLan Co., 1928.<br />

__--_. .--- .<br />

72<br />

.-<br />

#<br />

_.__e. c_ .--A____<br />

-- .---<br />

. .<br />

. . . .<br />

.,<br />

,<br />

, . .<br />

. *. .<br />

.<br />

.<br />

I<br />

.<br />

.<br />

.<br />

i<br />

;<br />

1.<br />

.


.<br />

.<br />

.<br />

.<br />

Parronality Tooting md Job Proficiency<br />

.?OHH D. RJtAFT<br />

US Amy Enlirted Bvaluatim Center \,<br />

Chairman Bridgeo, yellow’-‘Confetee.8,<br />

It ir a privilege to lead thir dirwrrion on perrmality teeting<br />

and Job perfomanca,<br />

A8 a point of departure in exploring waya of iacruring the prediction<br />

of roldiere’ performance, I vi11 give a goneral di8cuorion of<br />

the rarurch vhich Rayumd B. Cattell and hi8 u8ociatr8 at the Univorrlty<br />

of Illiaoir have conducted for mumming motlvetional rtrangth. Aleo, I<br />

will dircu88 8-e of the u8ea which wo might put Cattell’a re8march. and<br />

relrtod motivatimal ramarch, to in evaluating tho total percent of<br />

variance in the Job perbormanc~ of the individual rerviceman.<br />

Currently, the Army u8es A fob “aa8tary” teat to measure what the<br />

� oldler “can do.” It *JUU a rating rcrle to measure vhrt the roldier<br />

“vi11 do.” Unfortunately, even though years of hard lrr&or have barn<br />

attended by indurtriel prychologirtr in crying to refir.,+ the rating recli<br />

and to minimire it8 error8. it rtill ir a veak evalu8tfng iustrumnt.<br />

Individual8 differ in their ability to accur8tely parcaiw the cheractsrirticr<br />

of otherr. Sane of the factor8 relevant to thla differential<br />

perception are in the age, 8az, end prrronality of the parcaiver, la<br />

there charreteri8tic8 of the percoivad, and in the content erer8 in which<br />

the prrdic t ion8 are madu.<br />

Thr rating 8~~10’8 %ill do” � reee � re csotivetionol in content. They<br />

rhw the “drive” or “atntiment” f8ctorr o f t h e roldler’r performance. Hr,<br />

Shirkey, who rpoke before you A fw chute8 ago, recently did A factor<br />

analyrir 03 the Army’8 new CER (rating rcale). He found five factor@<br />

which can tentatively be celled: Militery Bearing I and II, Xilitaxy<br />

Laaderrhip Potential, Initiative, end T88k Uotivation. Hr. Shirkey wa8<br />

8urprired to find that the rupervirory portion of the trot correleted<br />

rather poorly vith the Leaderrhip Potential Pector of the CER. Although<br />

it io porrible that there rcrultr ‘Are partielly A condition of the prrtitular<br />

HOS analytad, it la evident th8t the CER la mesrurlng the mot;-vational<br />

arue concerned with vhat the coldior “will do.”<br />

Since mo8t ratiaga are rpurlour and do vary with thr rater, conrider<br />

how much better it would br if predictive mea8ure8 of tha motivational<br />

rtrength in there ereee could be nwde in order to give more<br />

objectivity and exactnerr. An cz8mple of the porrrble work that can<br />

be done along there line8 18 th8t of the rsrs8rch currently being<br />

conducted by the US Army Perronnel Rersarch Office on thr valtler a$aoelated<br />

with military cereer motiv8tion. Tlw rarearcher there heve<br />

73<br />

,-<br />

: ..o


-. _<br />

. . -<br />

:<br />

;<br />

-.----.-.--.e. .- -<br />

Identified the strengeh of six dimensions of motivational values or<br />

factors which are useful in predicting whether an officer or en epliated<br />

mm will plan on staying in the military service as a career. These<br />

factors which are tested by objective type teats ure support, conformity,<br />

recognition, independence, bcnevalence, snd leadership. ~1~0, they are<br />

currently conducting extensive research in developing devices which ~111<br />

predict cthich officers and enlisted uen will be the nest proficient<br />

dcring their &my careers.<br />

Cattell and his co-workers hsve spent the last fifteen yesra in<br />

trying to develop objective measures of motivational strength. They<br />

studied the amjor evaluative instruments produced by others, their devices,<br />

and the known general psychological principles in the areas of moeivatlon,<br />

learning theory, etc. (& example of the kind of principle referred to<br />

here is that of information. In general, i.e., after we discount the<br />

Influence of intelligence and general breadth of interest, a person knows<br />

more about what he is interested in, more about those courses of action<br />

to which he is comieeed, than he does about what he is not interested<br />

in.) Cattell developed m-me seventy-five devices of his own and poktulatcd<br />

8omf2 sixty new principles. Prom this moss of nnterial, he found<br />

through factor ana?ytic atudies,*seven basic raotivational ccmponents.<br />

ihree of these seem to correepond in content to the functfonr ascribed by<br />

Prcud to the Id, Ego, and Superego. The others have been named Physlological<br />

Interest, Repressed Cmplexes, Impulsivity or Urgency, and<br />

Persistence. These qualities can be ascribed in varying proportion to<br />

any drive. Second order facto- analysis of these motivational components<br />

resulted in two second order factors: (1) Integrated, or that which is<br />

essentially a conscious and experienced expression, and (2) Unintegrated,<br />

or that which is essentially unconscious and mainly wishfug and tension.<br />

The two most important working concepts of Cattell in his theoretical<br />

treatment of motivation have been termed “erg” and “seneiment,”<br />

respectively, for the constitutions1 � ud acquired patterns found<br />

opmationally as factors in dynamic measures by experiments. The term<br />

“erg” is used instead of drive because ehe latter term drags in all<br />

manner of clinical and other assumptions about “instincts” and so on,<br />

whcrers the ergic paeterns are experimentally demonstrable. How?Wr, In<br />

popular terms an erg is a drive or source of reactive energy directed<br />

toward a pareicular goal, nuch as fear, mating, assertiveness, aad so on.<br />

By contrast, a rentiment is an acquired aggregrate of attitudes built up<br />

by learning and social experience, but also, like an erg, a source of<br />

motivational energy and interest. Both ergs and senti;?rants, though<br />

essentially ccmnmn In form, are developed to different degreee in different<br />

people. Cattell and his co-workers found factor analytic evidence for some<br />

twenty motivational dimemions in a broad rampling of variables.<br />

-- _ . . _-._ -.-..___ -_- ___ _ _<br />

.<br />

74<br />

. . -.<br />

.<br />

I<br />

L<br />

.<br />

__.<br />

_.- -.-<br />

.<br />

.<br />

s


-. _- _. ._. ,. _ -- _ ______-________ ---..--- ..-_--_--___. .--..- ----.. -... -- .<br />

. *<br />

. .<br />

& .’<br />

.<br />

_ _ -----.- --A- __.__.___” ____ _. ,__- _-_..... ..-. -.._<br />

S~aa of the major crgic patterns or dynamic factors for which Cattell<br />

found evidence through his research are: Escape, with associated emotion<br />

of fear; Hating, with associated emotion of sex; Seif-assertion, with<br />

associated emotion of pride; Gregariousness, with associated emotion of<br />

loneliness; Appeal, with assocfated emotion of despair; Exploration,<br />

with associated emotfon of curiosity; Rest seeking, with associated<br />

emotion of sleepiness or fatfgue; etc. Some of the major sentiment<br />

patterns or dyuamlc factors for which he found evidence are: Self;<br />

Superego; Career (levela of aspiration); Sports and games; Mechanical<br />

Interest; etc.<br />

Cattell and his assocfates then developed the experimental HotLvational<br />

Analysis Test to measure the tension operative in the ergs and<br />

sentiments that prior to nov was assessed only by rough means. Ten<br />

dynamic factors were chosen which were felt, as a result of careful correlational<br />

and factor analytic research, would be of greatest value to<br />

the test users as being reprerentative, and comprehensive in coverage,<br />

of adult motivations.<br />

In this test, Cattell mearured the strength of each of these ten<br />

dynamic factors by using forced-choice type of attitude-interest questions.<br />

The particular attitude-interests he used were carefully chosen because<br />

they were found to be substantially -related to and therefore best suited<br />

to represent the factors concerned. For example, he found the folloving<br />

attitudes and their motlvatfonal paradigms for the dynamic factor of<br />

self-sentiment: ,<br />

-.<br />

1. Good reputation - - “1 want to maintain a good reputation<br />

and a camon respect in my commmity.”<br />

2. Norm1 sex - - - - “1 want a norms1 socially acceptable<br />

relation to a person of the opposfte<br />

sex.”<br />

3. .Look after family - “I want to look after my family so that<br />

it reaches approved social standards .‘I<br />

4. Proficient career - “I want to be proficient in my career *”<br />

5. Control impulses - “I want to keep my impulses under sufficient<br />

and proper control.”<br />

6. Self respect - - - “I want never to damage my sense of<br />

self respect.”<br />

The motivatioru!l strength of these attitude-interest paradigms for<br />

each dynamic factor vere measured by using four different forced-choice<br />

type devices or subtests. These subtests, which are a little unusual<br />

in nature, are called: (1) Uses; (2) Estimates; (3) Paired Words; and<br />

(4) Information. They are illustrated in the folloving chart.<br />

75


�<br />

�����<br />

:<br />

���<br />

�☺�<br />

.-<br />

.*<br />

.<br />

- . -_ _ . __ ma - . -. -. . _. _ __ _<br />

_.I ,<br />

.i<br />

2.<br />

-:<br />

_. .’<br />

. -.<br />

.<br />

�<br />

�<br />

��<br />

�� �<br />

� �<br />

�<br />

� �<br />

--__~. _-. -. --. ._<br />

�<br />

�<br />

Subtests (Rxamples~<br />

1. “Ends-for-means” (Projection) or “Uses’.’<br />

A free afternoon could beet be spent:<br />

a. enjoying the out-of-doors<br />

b. earning overtime<br />

2. “Autism’@ or ~*Esthates”<br />

How many years does it take to make a secretary efficient?<br />

a. 2 b. 4 c. 6 . d. 8<br />

3. “Ready <strong>Association</strong>” or “Paired words”<br />

Stamps<br />

Collect<br />

Evidence<br />

4. “Mean-end Knowledge” or “Info~tion”<br />

Who was the first president of the United States?<br />

a. Lincoln B. Jefferson C. Washington D. Roosevelt<br />

The scores from the four su3tests are combined into two groups for<br />

each factor to give separate strength measures for: (1) the Integrated<br />

(or conscious) (Information plus Paired Words) component, and, (2) the<br />

Unintegrated (or unconscious)(Estfmates plus Uaea) component. Thus, the<br />

test user can either use two scores for each of the ten dynamic dimensions<br />

in the N. A. T., or Integrated and Unintegrated scores can be added to<br />

give a single total interest score for each of the ten dynamic dfmenslone.<br />

The final score on each of the ten dynamic structure factorcr i8<br />

actually a sum both Over four devices, or methods of measurement as just<br />

described, and aleo Over distinct attitudes which are known to constitute<br />

good representatives of the dynamic trait. There are a few other scales<br />

developed with this test. Perhaps the most interesting one ist the<br />

Conflict Scale. This scale shows the amount of conflict between the<br />

Integrated and Unintegrated scales.<br />

Cattell reported the scores in stens (standard scores) for each of<br />

these scales. The statistical procedure used in arriving at these scores<br />

is ipsative in nature. One can learn the total strength of each of these<br />

drives in the individual being tested and be able to make diagnostic predictions<br />

from them.<br />

if -. .-._. ..-.. _~__. _-- ---~--.. -._-_ -<br />

76<br />

. . .<br />

4


.<br />

.<br />

. .<br />

,’ . .<br />

NW we rhould recognize that what Cattell is measuring here a8 a<br />

“drive rtrength” is actually mO8t accurately rtferred to ae an ergic<br />

ten8iOn 1eVel. That i8 to 8ay, it i8 not 8omc donstituticnnl strength<br />

of that drive, but the end result, in actual motivational 8trcngth of<br />

four or more influencer. There are: (1) con8titutional endowment, (2)<br />

early experience8 (imprinting, reprerzioc), (3) current degree of environmental<br />

8timulation, and (4) current reduction of tcnzion by satisfactions<br />

and degree8 of recurrent need gratification existing in the prerent<br />

environment. Theae vary with time.<br />

The Motivational Analysis Test is etill an experimental tert.<br />

Hwevtr, a8 more rerearch i8 conducted on it, it8 theoretical predictive<br />

ability and ��� � cutal predictive ability a8 derived fran preliminary<br />

reraarch rerulta will increase.<br />

In predicting what an adult will do, it 18 evidently quite a8<br />

important to knw the motivetion available frop these acquired rentimentr<br />

a8 from hi8 basic ergs. For urimple, the evaluation of a man’8 potential<br />

In a career will need to include mea8ure8 of the rtrength of hi8 intertrt<br />

in a career a8 8uch, 8180 of hi8 degree of concern about hi8 general<br />

reputation and self-concept, and ezp&ially of the level of basic dependability,<br />

implied by his level of superego development.<br />

The 8tati8ticalpredictionzwhich rezearch has 80 far obtained<br />

against life criteria have borne out these poychological interpretations<br />

of 8core8. Rerearch ha8 shown that dynamic structure trait rtrengthe<br />

add to the prediction of scholaetic achievement over and above what ir<br />

predictable from ability and general perronality traits. John Butcher,<br />

one of Cattell’s a88ociate8, found that other variable8 being controlled.<br />

those students who are high in ruperego ��� � ��� high � chiever8, there<br />

who are low on radi8m are high � chiever8, those that are high on<br />

arrertivene8s are high � chiever8: but he found no relationrhip with<br />

8elf rentimcnt. Aleo, he found that the career sentiment mU8t be modifiedI<br />

for the particular setting.<br />

Mr. Claude Bridge8 talked with Dr. Arthur Swency a few week8 ago.<br />

Dr. Swenay, who ie an � 88ociate of Cattell, 8aid that the preliminary<br />

reoearch rhwr that general motivational factorz could contribute a<br />

great deal a8 predictor8 of achievement in 8ptCifiC occupational rpecialities.<br />

He raid that it would probably be neceroary for u8 to develop<br />

rpecific inrtrumentr for group8 of occupational rpecialitier. (fii8 got8<br />

in with another te8t that Cattell has developed called the Vocational<br />

Interrt Heawre, which ir built on exactly the came principle8 a8 the<br />

Motivation81 Analysis Test only it iz geared toward predicting in 8pecifi.c<br />

occupational groupinga. The Vocational lnttrt8t Hearure ir being validated<br />

againzt the Kuder Preference Record.)<br />

.<br />

77<br />

-. -<br />

- :r ---<br />

i .


’ ._ ,_<br />

1 _ +<br />

_ _ . . .<br />

.<br />

a<br />

--.<br />

s<br />

.<br />

. .-. . - . _<br />

. . .<br />

/ . . .u<br />

/ . .- - .- . . ~. :.. __..----.----._- 11.11_--___ _.<br />

*<br />

---- - __------.<br />

Cattell says that when the results of the Motivational hnale<br />

Test are applied to clinical and industrial work,ritarion re%tiona<br />

oreequally significant and comprehensible in terns of the concept<br />

validities of the measures as they are of school achievement. Since<br />

Cettell is currently publishing the predictive details of most of his<br />

industrial rcseareh, it will be some months before these results ara<br />

1 available. .<br />

I feel that Cattell’s work is vary important to us as he has<br />

broken through several barriers. First, by carefully reviewing the<br />

previous motivational materials and those of his own, he wae able to<br />

corns up with motivational factors which operate more or lesr in all<br />

adulto. Some of these factors could be very valuable to us in OUT<br />

efforts to evaluate the total, on-the-job effectfveuers of each soldier.<br />

For example, by utilizing test results on the relative tension levels<br />

of ~cnae of these factors, we could after careful etudy of their rigniffcance,<br />

utilize the test results to identify the cause of a soldier not<br />

working up to his pest effectiveness.<br />

Second, Cattell developed test procedurea whereby the variable<br />

being measured is 80 disguised that the subject usually has no desire<br />

to distort hir response (e.g., what is obstensibly an ability tert) OT<br />

else would be incapable of doing so in any ays:ematic way because he<br />

ie unaware of a “desirable response.” Cattell has gotten away from<br />

the drawbacks of deliberate faking, of personal illusion, snd of ruperficielity<br />

of Easurements which have vitiated against mast “projective,”<br />

“preference ,‘I and “opinionsire” methods employed In interest batteries.<br />

Since Cattell hes shown that a person’s rrotivational makeup ia baaed<br />

on both Integrated (conscious) and Unintegrated (unconscious) components,<br />

we can see that by simply asking a person how he foels about something<br />

this cannot possibly be sufficient to tell his rrPtivationa1 ten.sion in<br />

the requested area.<br />

Third, Cattell utilized new scoring procedures, ipsativa in nature,<br />

whereby he could learn the total strength of each of these drives in the<br />

individual and be able to make predictions from them.<br />

Although Cattell’s Motivational analysis Test ia rtfll a research<br />

instrument, it is a very major step forward in mJtivationa1 strength<br />

measurement. This test covers a person’s Interests, drives, and the<br />

rtrengths of his sentiment and value systems. It increaser the reliability<br />

of conclusions regarding perromlity dynamics and helps to<br />

locate areas of fixation and conflict.<br />

The test concentrates on ten prychologically meaningful unitary<br />

olotivation systems, established by comprehensive and objective factor<br />

analytic research. Also, it uses nawly validated objective test devices<br />

for measuring these interest strengths.<br />

78<br />

. . -. _... - _.... -.----.-____ .-.- -- -_--_--<br />

. .<br />

.<br />

.<br />

.<br />

.<br />

.<br />

.<br />

.<br />


_ . - . .._ - -_-..-.-- _-_- ._,__X_____.___r_C_ _ --_ .._.-.. - .<br />

- _ __ .-_. - _ -~. .__ ..__ .-_ -. - _____. - _- _ _. .- _ _ _<br />

.<br />

. .<br />

.<br />

_.- ---....-I- --...- -~..-.‘---.------.-. C....‘~e-.C.---F --*- -..s.*---1m-*r~- --...-,.. -..ea.,I .,,_.__. _.‘_,” ,_ I ,, ..<br />

By centering on the main mtfvvation nyetenm, ns demonstrated to<br />

be consistently present in the typlcal indFvFdua1 of our culture, rather<br />

than on “special purpose,“ arbitary conglomerations, .eubJectivelyconceivad<br />

interest divisions, or iamginary “drives,” the test design<br />

recognizes the whole adult.<br />

It may be possible that we can use the research findings behind<br />

these tests. We can develop our own motivational fastor tests from<br />

the great fund of current, veil-founded, and documented research.<br />

Since we have large populations (serviceman) avaiiable to us, we can<br />

validate the tests rather easily. ~160, with our quite adequate computer<br />

machinery to work with, we can easily do the required statistical work.<br />

The tests which we could develop can be wed in three areas: (1) to<br />

rupplement or replace parts of our rating wales, (2) to set up a criterion<br />

for validating and improving ‘*supervisory and cognitive” portions of our<br />

present teata, and (3) to set UP a criterion for validating our rating<br />

rcales against.<br />

.<br />

Conclusions<br />

When we look back at Chain.ln’s Bridges’ chart here, and see the<br />

large percent of variance that is not covered by general *‘achfevewnt”<br />

type job-knowledge tests and rating scales; and when we see in the<br />

one case of Mr. Shirkey’s Factor Analysis of, the Army’s new rating scale,<br />

that our supervisory section of the Army’s Job mastery test dt-.e not load<br />

highly vith *he factor of Leadership Potential; and when we remember that<br />

the rating scale is influenced by many factora of the rater, ratee, and<br />

the content setting, perhaps this ic a major area which we should explore.<br />

Chairman Bridges and I would like to nou open the session to a general<br />

discussion of personality testing and Job performance.<br />

.<br />

79<br />

.<br />

” *<br />

.<br />

!<br />

I<br />

I<br />

I


,’<br />

�� ���� � �<br />

����<br />

s<br />

. --.<br />

� � � ��� ���� �<br />

. *<br />

-. .<br />

. . I<br />

.- .A-..-. ..--7 -------.-_<br />

- : - . -. ._^_...<br />

� �� � ��� � �� �� � � �� �<br />

Concluding Cameente<br />

CLAUDE F’. BRIDGES<br />

US Army Enlietad Eveluetion Canter<br />

___ _--._-_ .-.- . -_ *.---. . -- -.-. _. -_---.--.<br />

-. ._--<br />

After tha etimuletfng preperetory presentations of Cenarel Z8i8,<br />

Ceptain Reyee, end Dr. Edgarton, the question � rieee ‘What have we<br />

� ccanpliehad in this eympoeiu~?” Cur axpectetion wee not thet WI would<br />

imeadietaly make eny breakthroughs into the outer specs of tasting. We<br />

did hope to go beyond mora axchenga of procedures end obtain acme etimu-<br />

1Ation end nascent ideee having potential for meking inroede into tha<br />

currently 1nadequetaly maaeureble percentega of job proficiency. I<br />

believe wa heve � ccanpliehad this purpoea. Three d1ecueeione hevc provided<br />

us with idaee, possible new � pproechce, or naw epplicetione of<br />

aetabliehad techniques thet are worthy of further axploretion.<br />

Dr. Zaccerie mey strike some treditionel1et8 as a bit of 8 maverick<br />

ia some of his ueegae of terminology, but the points mada in tha paper<br />

by him end Dr. Kerp marit serious coneidaretion. True psychological<br />

8nelyece of jobs, epccificnlly for � aeeurament and treining purpoeeel<br />

and the cfftctiva detenninetion of functions1 job requirements and<br />

� tenderde should prova to be one of our most fruitful � terting ?ointe.<br />

As Cronbech (1960) etetee, “The most important requiranent fot valid<br />

� eec-Jemant le...8 cleer undcretending of the psychologies1 rcc,uiremente<br />

of the criterion teek.” In this connection, severe1 recent publications<br />

by Dr. Robert N. Cegnc end Dr. Robert B. Hfller art pregnant with poeeibllitiee.<br />

The pepare by Hr. Urry end Hr. Shirkay indiceta not only that con-<br />

� idcrabla improvament of even good taste ten still result from further<br />

polishing, but � leo point out some new applications of etetFeticel<br />

� nelyeie end control techniques to npprecieble incraeea in test velidity.<br />

..- * AS expected, Lieutenant Longo’e peper provides eavarel significant<br />

contributions. The Nevel Air <strong>Technical</strong> Treining � teff, under Dr. Meyo,<br />

,<br />

for many yeere has bean conducting axteneivc reecerch on tha prediction<br />

of job proficiency end on � ctuel versus predicted � chievament. This<br />

pepar werrente some edditionel ccuxaente end I would like to meka sane<br />

further suggestions on the use of an indax beeed upon the relationship<br />

of (1) ectuel � chievament, es mteeured by a proficiency test, and (2)<br />

axpectcd � chitvament, as predicted by en � ppropri&te � ptituda test.<br />

Lt Longo pointed out, that such 8 “diecrepency score” or, as I prefar<br />

to cell it, “effcctfveneee indax,” reflects the composite affect of<br />

several verieblee including both error verienca end � ll verience not<br />

in cOrnmOn between the two varieblee. Included in these influencing<br />

vrrieblce era attitudes, work hebite, locelly unique job eituetion<br />

varieblee, interection effects, end motivetion-in other words, the<br />

affectiveneee with which the men � ppliee his eptltudee, ekil.8, and<br />

knowladgce to achievement in 8 specific � ituetion.<br />

80<br />

.<br />

. .


.<br />

-. . . . -- .--- ..-_ .--- . .--. - -- ..- -- ._ _.. - _.. -._<br />

.<br />

. .<br />

Two usea are proposed. First, careful study and anAlyses of the<br />

differences in characteristics of “over-achievers” and “under-achievera,”<br />

using what Dr. Helvin R. Marks calls “the off-quadrant upproach,” ahould<br />

help identify variables Chat need to be covered better by the tests.<br />

Second, is an adaptation of the usual procedure that may add significantly<br />

to the multiple Correlations.<br />

In his doctoral dissertation, Reverend Brother James C. Bates, F.S.C.<br />

Consultant to the Provencal, Christian Brothers of Ireland, Canadian<br />

Provence, recently analyzed sole of the concomitants of the “learning<br />

effectiveneso” of 381 male students in a liberal arts college. While<br />

checking out some of my hypotheses, he produced evidence that an<br />

appropriately developed effectiveness index combined with an achievement<br />

test and an aptitude test does yield a significantly greater multiple<br />

correlation with academic marks than using only the two variables from<br />

which it was derived. Naturally, the addition to a multiple regresoion<br />

equation of a variable which ir a linear function of variables already<br />

included cAnnot increase the multiple correlation coefficient. Even<br />

using the ratio of the two should Add little. The correlation between<br />

the two types of indexes USUALLY approaches unity. ,<br />

However, in one part of Brother Bates’s study, the effectiveness<br />

index formula was developed from a total entering Freshman group (381)<br />

and then applied to the 274 students who graduated. When combined with<br />

the other predictor variables, these effectiveness scores increased ..he<br />

multiple correlation with first year college marks .OS points, which was<br />

significant at the .OOl level. Furthermore, the inclusion of this total,freshman-group-bnsed<br />

effectiveness index in the regression equation to<br />

predict average of allcollegeEarks for four years yielded the remarkable<br />

increase of .13 in the R. The canposfte of averc.:e high school marks,<br />

Essential High School Content Battery (the achievement test used in<br />

developing the effectiveness index) and the American Council Psychologic’s<br />

Examination (the aptitude test used),yielded an R of .56, which WAS<br />

raised to .69 by the addition of this type of effectiveness index.<br />

Incidentally, due to the hfgh reliability of the tests used to develop thlcm,<br />

these effectiveness indexes had A .93 reliability coefficient.<br />

This same procedure possibly could be adapted to the military situation.<br />

It should prove useful and relatively easy to develop in connection with<br />

successive levels of training. For the Army, an effectiveness index for<br />

a given F@S based upon the best combination of aptitude measures vereu8 t.hs<br />

Enlisted M0.S Evaluation Test scores and derived from a total Ho.9 skill<br />

level population might increase significal:tly the prediction of the future<br />

proficiency and promotability for soldiers in one pay grade or other<br />

identifiable auhgroup. Other possibilities for meaningful rubgrapa<br />

might be devised and checked out. Verification of the applicability to<br />

OCCupationAl dpecialties of the findings in Brother Bates’s dissertation<br />

offers the possibility of yielding A simple inexpensive way of significantly<br />

increasing validity. A true moderator VAriAble seems to be involved here.<br />

81<br />

;


. . .-.<br />

._ .<br />

.<br />

-. -.<br />

Anyone investigating this area further will find several references<br />

helpful. A publication edited by Drs. Dubois and Hackett includer reports<br />

by Dr. tlayo and Dr. Froelfsh referred to by Lt Longo. Dr. Harris’s book<br />

contains an excellent presentation of soma of the complex statistfcal<br />

problems and theory involved in this area of research. Dr. Thorndike’a<br />

little monograph examines closely the concepts and methodological problems<br />

and makes specific suggestions for sound research studies in thia<br />

rather tricky area.<br />

The lines of approach reported in Mr. Kraft's paper offer some seminal<br />

points of departure. III fact,the greatest hope for a major increase in<br />

percentage of job proficiency measured probably lies in obtaining more<br />

reliable and precise measures of the aspects of personality related to<br />

job proficiency--better ways of measuring the factors comprising the<br />

“will do” of enlisted personnel as identified by psychological job<br />

analyses and research. There is a good indication of the soundness of<br />

this conclusion, On 3 August 1964 in a letter on “Revision of Army<br />

Aptitude Area System”, Headquarters, US Continental Army Command,<br />

stated as follows:<br />

. . . Numerous school-conducted studies have shown that age,<br />

previous education; previous civilian and Army experience \<br />

of students frequently correlate hi&her with course performance<br />

than do ACB test scores. It is recommended that<br />

appropriate statistical precaution be taken by USABRO to<br />

Insure that factors other than test results do not contaminate<br />

the validation.<br />

“Evidence presented at an April 1964 USCCMRC Basic<br />

Electronics Conference, plus observations of numerous<br />

key personnel throughout the USCONARC school system, indicate<br />

clearly that subjective factors, such as attitude,<br />

desire to study, perseverance and other non-intellectual<br />

attributes of USCONARC school students have as much or more<br />

influence on courses performance of these students than<br />

“aptitude” as measured by classification test battery such<br />

as currently employed or contemplated by USAPRO. To the<br />

extent that inputs to USCOMRC school and training center<br />

courses are governed by tools thst take into account only<br />

those cheracterfstics readily measurable and ignore more<br />

significant - though admittedly subjective - factors,<br />

there will be less than optimum regulation of manpower<br />

flow through the Army training system. USAPRO is urged<br />

to consider the development of measures of attitude,<br />

motivation and desire to learn as an integral part of its<br />

program to revise the Army Aptitude Area System.”<br />

.<br />

02<br />

c<br />

.<br />

.


.<br />

.<br />

.<br />

In their reply the US Army Personnel Research Office stated:<br />

..* the background data reported will include not only<br />

ACB scores but also such factors as age, cducatkon,<br />

cfvilfan experience, etc. These variables wtLK be included<br />

in the analysis with the experimental test data.<br />

“The non-cognitive factors referred to *.. have indeed<br />

been demonstrated to be important in training performance,<br />

and in jcb performance as well. In this regard<br />

it Is worth noting tbct the erperinen:al battery<br />

contains several non-aptitude type measures, designed<br />

to evaluate those enduring interests, attitudes, rind<br />

other personal characteristics which help determine<br />

what a man - vi11 - - do - rathtr than cnly what he can - do. -<br />

One goal in the revised ACB is to provide a<br />

Classification Inventory which vi11 measure personal<br />

characteristics predictive of performance in occupatfonai<br />

areas such as electronics, mechanical maintenance,<br />

clerical-admlnfstrative, and other areas, just as<br />

the present CI measures personal factors which predicted<br />

combat performance in the Korean War. 2<br />

-addition - to these enduring --s--Y characteristics. - - -a__- however *<br />

there are factorof my.--_ a~otivstion which .e-are prfmnrlQ<br />

-situatfocel. - The classification br,ttery cannot predfct<br />

these, but this r Isearch program must take Lnto<br />

account the effects of such factors on the flndinge.<br />

It will be very helpful to the USAPRO scfentieta if,<br />

in conjunction with thfe program, the training installationr<br />

can contribute inrlghtr end data on such<br />

factors.”<br />

In this theoretical sympooio?n M have been able to little more thrrr<br />

touch on possible approaches to more effectively measuring some of the<br />

eight types of characteristic8 listed in the tntroduction. Thin afternoon<br />

a fev more possible approaches wfll be presented, Some should<br />

prove productive, and all should be stimulating. For example, in some<br />

occuprtfonal specialties even gross methods of controllfng item readability<br />

should yield prcfitable improvements especiallqr in tests for occupational<br />

specialties in which the level of reading ability of incumbents f~ a<br />

significant contributor to invalid variance of Hcorca. These improvements<br />

should be even greater for such tests i.f three 8teps can be accomplished.<br />

First, If a more precise item readability index can be developed instead<br />

of using adaptations of the gross sampling methods that are adequate for<br />

masfies of regular prose. Second, if a practical method can be developed.<br />

for correcting this index for special technical terminology characteristic<br />

and counon inthe occupational specialty. Third, if some means of determining<br />

the distribution of reading ability in each specfalty ia<br />

feasible.<br />

83<br />

---- .-. ..- .--_ -_8_-.-.-- ..- - ._ -.-. -..- ._ .__.(._<br />

.<br />

*<br />

k


:/<br />

. - . -_ _* .-- - -.<br />

-- . .<br />

. . ’<br />

. \<br />

./:.<br />

‘,<br />

.\<br />

-. .<br />

. . -<br />

.<br />

.<br />

--.---_ -- .-._. ---- I- ._- . _ _ - __--------.._<br />

To mention one more forthcoming approach, the “Performance Check Test”<br />

concept may offer a practical aad logically 8&nd way of effectively<br />

measuring motor skills in tho8e occ’upational apeclaltiea for vhich such<br />

measure8 are important and for which the motor skills meet the required<br />

criteria.<br />

Haking what Dr. Edgerton identified a8 genuine change8 in testing<br />

obviouely will require much effort, ingenuity, and a fortuitous concourtit<br />

of circumstances. .<br />

T &us confident that at leaet 8ome of the profe88ional 8taff fras<br />

each of the military services and in other research agencier will<br />

� ggreeeivaly pursue any inrightr that may be stimulated by these 8ympo8ia.<br />

We have much at 8takC and many of the 8aae problem8 in coamon a8 indicated<br />

both by Hr. Johnson’8 mmxnary analysi8 and cauparf8on of the programs<br />

of each agency and by your answer8 to the program planning qutstionnairer.<br />

These tend to emphasize the coaxson lnterert in exploring wnyr of obtaining<br />

marked increase in the percentage of overlap between job factorc and<br />

the evaluation8 of enlieted ptrronatl. You relectcd a8 tha theme for<br />

this conference, “Xncrea8ing the tleaauring Efficiency of Evaluation<br />

Instrumento.” Your rerponse8 to the program planning questionnaire8 indicate<br />

that, regardlee of the current emphasis in the evaluation program<br />

of the five services, an inwcdiatt goal of each service 18 the objective<br />

eValUatlon of as many significant factor8 iu Currtnt job Mattry a8 ie<br />

practical. For all uae8 of the respective programr., apprairing the<br />

Current ltvtl of total job proficiency is importsnt. Por promotion8, the<br />

ft.ture level of job proficiency is the important intermediate criterion.<br />

Of course, for all of the servicer the ultimte crittr1c.i is vhather or<br />

not we win any military action and, for the individual enlisted man, is<br />

hov well he function8 in hi8 a8signtd job during such action.<br />

� �<br />

.<br />

� �<br />

I .<br />

- .- _ _ _ - .-..- - ---_.- . _-_-- ---<br />

.<br />

.<br />

.<br />

I .<br />

. .<br />

, .<br />

.<br />

�<br />

.<br />

.<br />

.<br />

.<br />

.<br />

.<br />

b<br />

.


.<br />

- _ _<br />

-<br />

.<br />

- - . .* _<br />

.<br />

.<br />

.<br />

-M1_1-_---.--“--- --I<br />

><br />

Reference8<br />

.<br />

-. .-<br />

.<br />

_... ._ _.- . . ..--.. - - -<br />

B&e., J. C. An inalyrir --I--.<br />

of some of tha concomitant8 af learning<br />

afftctfve~rs. PhD dferertation, St. John’8 Univerrity, New<br />

York, 1965.<br />

Cronbach, L . J. Elrentlalr of prpcholoRicr1 testing (2nd ad.).<br />

Nev York: Harper, 1960, 589.<br />

Dubofr, P. H. & Hackett, E. V. (edr.) Tb8 &8Ure¶IMIt and evaluation of<br />

over-and under-•chlsveznant. St. Gi8: Waohingto~nlverrlty, I.%2.<br />

time, R. H. The � cquirltfon of knowledge. P8whol Rev, 1962, 69, 41,,<br />

355-36s.<br />

Sagne, R. H., Major, J. R., Carrttnr, Helen L., & Paradise, N. E.<br />

Factor8 in acquiring knwltdge of a mathcmotfcal tark. P8ychol<br />

Monogr, 1962, 76, No. 7 (whole No. 526).<br />

Gagne, R, H. & Paradire, N. E., Abilititr and lernfng ret8 in knowledge<br />

8cquirition. Plychol Mono& 1061, 75, No. 14 (Whole No. 518).<br />

. - . _ _<br />

Harrla, C. W. (cd.). Problems & mearurlng Mchrnge. a d i s o n : T h 8 Unlvrrrfty<br />

of Wirconrin Prt88, 1963.<br />

Xillsr, R. B. Handbook 0” training & Jraining equipment dt8iRn.<br />

Wright-Patterron Air Force Base, Ohlo: Wright Air Development Cmnttr.<br />

Technicel <strong>Report</strong>, 53-136, 1953.<br />

HilIar, K. 8. A method for man-machine taok ana?..ysir. wright-Pattsrroo<br />

_I-<br />

Air Force ~8~i~Wrlght Air Development Center, <strong>Technical</strong><br />

<strong>Report</strong>, S3-137, 1953.<br />

Miller, R. B. Soma Pittrburglh:<br />

workin& COnCtpt8 of ryrtem8 analymir.<br />

American Ixtute for Rt8earch. 1954.<br />

Hilhr, R. B. Tark and part-task trelntrr. Wrfght-Patterron Aft Force<br />

- -<br />

Bole, Ohio: Wright Air Development Center. Ttchnlcal <strong>Report</strong> 60-460,<br />

ASTIA No. AD 245652, 1960.<br />

Miller, R. B. 6 Van Cott, H. P. The dtttmfnation of knowledge content<br />

-me<br />

for complex mm-machine jabr.-Pittrburgh: Ams&an Inrtttutt for<br />

Search, 1955.<br />

. Thorndike, R. L. The concept8 of over-and under-achievement. New York:<br />

Bureau of Pub-ationr, Tezherr College. Columbia Unlverrlty, 1963.<br />

85<br />

, !<br />

k;<br />

.<br />

. 1<br />

!<br />

/<br />

,


-<br />

._<br />

.<br />

. .<br />

. .<br />

\<br />

.’<br />

.-<br />

.--<br />

.:.<br />

‘.<br />

-<br />

-. _-<br />

.<br />

.-<br />

_<br />

_..<br />

-.<br />

t-.. .:.<br />

-<br />

_.<br />

. ;<br />

-. -7<br />

- .zsI,:<br />

- :<br />

---.i<br />

: -.+,<br />

.-<br />

.<br />

. -w.<br />

. *<br />

-. .<br />

. . m .<br />

.<br />

.<br />

---- -.- ..--.. ._ _. __ -__. _.-. -__-------.-- ._i-_.. - .- ---<br />

THEOREnGAhfYlsl~OSlUfali-LEROYJOHI1SlOn,CHAI~AR<br />

USWAVALEXAIBI#IIGGE#TEB<br />

Relationship Between Raadability and Validity<br />

of 140s Evaluation Tests<br />

.TCXW S. BRAND<br />

US Army Enlisted Evaluation Center<br />

Cronbach states‘in his book Essentiala of Psychological <strong>Testing</strong>:<br />

“Tests having the same ‘content’ may measure different abilities because<br />

of variables associated with itam form. Reading ability, for example,<br />

affects scores on almost all achievement tests. A valid measure of knowledge<br />

is not obtained if a pearson who knws a fact misses an item about<br />

it because of verbal difficulties.” Cronbach goes on to ahow the superior<br />

validity of picture items over verbal items.<br />

A number of methods have been proposed for measuring the readability<br />

or reading difficulty level of written matter. The principal ones investigated<br />

by this offfce are those of Gunning, Flesch, and the Farr-Jenkine-<br />

Paterson modifi:ation of the Flesch method. These three methods, which<br />

are essentially variants of the same basic principles, were initially<br />

applied to several samples of written matter. It was concluded that the<br />

results obtained from the three methods were essentially the same. The<br />

Gunning method, however, has the advantage of being simpler, and was<br />

therefore adopted as a method for the measurement of readability.<br />

This psychologist is only aware of two papere dealing with the<br />

readability of tests. Neither of these present methods for determining<br />

the readability of individual test itema.<br />

Long words and long sentences have baen shown to be the principle<br />

determiners of readability. The three methods mentioned above all use<br />

average sentence length as one determiner of reading difficulty. In<br />

determining average sentence length, independent clauses are counted<br />

as sentencea.<br />

Flesch counts all syllables to determine the average word length.<br />

Farr-Jenkins-Patterson suggest that the same information may be obtained<br />

more easily by counting one-syllable words. Gunning counts words vith<br />

three or more syllables to determine percentage of “hard” words.<br />

--_-_. -.. .._ _ - - -<br />

.<br />

--s-- --<br />

. . . . :<br />

86<br />

.<br />

.<br />

_--___-. ..-._. --<br />


-.<br />

. :a.<br />

.\<br />

-.<br />

-I-<br />

.’<br />

-.., 7<br />

-a!<br />

i<br />

-.-<br />

.- /-<br />

-, *-<br />

.‘.<br />

-..<br />

,’<br />

-<br />

..- ._<br />

.-<br />

. .<br />

/’<br />

-: ,-<br />

-y _<br />

--<br />

$11<br />

- .-<br />

-<br />

- -<br />

--___<br />

.-_<br />

.<br />

.<br />

In counting three syllable wOrd6, Gunning makea the following<br />

exccptlona:<br />

verb forms that are made three syllable8 by the addition<br />

of -CI pr med. For example: reducte, invested<br />

proper nouns or word8 that are capitalized<br />

combinations of simple words, like bookkeeper<br />

t&nning’S Fog Count, as he Call8 It, i8 found by adding average<br />

sentence length and percentage of herd words, end multiplying the resulting<br />

am by .4. The Fog Count is a grade equivalent. Twelve equals high school<br />

graduate, 16 equals college graduate, etc. I have celled the acore<br />

resulting fran the application of Gunning’8 method a “Readability Index”‘,<br />

or RI.<br />

The fOrUtUla8 develdped by Gunning end Flesch vere validated against<br />

the McCall-Crabbs retding tests which were etandardized against ext,enslve<br />

student population8. This critcriOn is admittedly not a good one for<br />

adult reading materiels, but it vaa the only one available for that<br />

purpose 0<br />

Application of Gunn’ng’s Fog Index to MOS Evaluation Teat Ire.2<br />

Plesch end Gunning recommend taking loo-word samples of long articles<br />

to determine reading difficulty. Counting only one alternative to avoid<br />

excts8ive repetition, the average test iten has about 20 words. This 01<br />

admittedly a small sample. However, much Useful information may be gained<br />

if a readability index can be determined for the individual item. An<br />

echlevement-type item will contain one or more complete sentences to<br />

determine average sentence length, end the number of 3-eylleble word8<br />

in en item can be counted to determine percentage of “hard” words.<br />

In counting 3-syllable worde, decieione mU8t be made about counting<br />

number8 and eerles of lettera, or numbers and letters, which occur feirlly<br />

frequently in MOS evaluation tests. In this respect, it is believed<br />

that Gunning’8 ryrtem is more adapteblc to determining the readability<br />

. of HCS evaluation test items then that of Flesch.<br />

Determination of Validity (ric)<br />

A criterion of job performance consisting of 3 peer ratings per<br />

subject on an overall performance scale hae been obtained by this Center<br />

for samples of EM in selected NOS. These ratings vere obtained in field<br />

trip8 to selected military installations. Item validity (ric) was<br />

obtained by the point bi8eri81, or Pearson correlation between item and<br />

criterion.<br />

. ,<br />

.<br />

87<br />

.<br />

.


“.<br />

.I<br />

-..-<br />

--;-<br />

_, ‘5<br />

.’ ,<br />

I<br />

. --_--<br />

--- - -*. --<br />

. . -<br />

^<br />

.<br />

.<br />

a<br />

.<br />

- -<br />

._ _. .,-. _ - _ _..- .___ . _.... . .._<br />

Analysis<br />

A bivariate distribution ~58 prepared betwoen readability level,<br />

shown on the horizontal axis in intervalt of 3, and ric, shown on the<br />

vertical axis, in intervals of .lO.<br />

Correlational analysis to detect nonlinear r&%tfanshipr of one<br />

sample produced nonsignificant results.<br />

The ric was cmputed for each column of the scatterplot. E5ch<br />

column represent5 different levels of reEdability ic intervals of 2,<br />

e.g., 2-3, 4-S, etc., through 22-23, 24-25, etc.<br />

The items of the teat were divided in several way8 on the basis of<br />

readability level. The-tic was computed for items above a certain readability<br />

level, and the ric was computed for iteXM below this level. If<br />

the former is lower than the latter the sigcificance of the difference<br />

was tested. A one-tailed te8t was used. Degrees of freedom for each<br />

ric vas taken aa the number of i:ezns included in computing the ric times<br />

N-3 of the validation sample. In other word5, each item contributes<br />

N-3 degree5 of freedom. .<br />

The critixal readability level WBR tnken oa that diviafon point which<br />

yields the largest s score or critical ratio betaeen the ricls obtained<br />

for the two parts of the test.<br />

The yic’s so obtained were also tested for significant deviation<br />

from zero.<br />

To dete,rcsults have been obtained in four MOS. Results for two<br />

of these are shown in the following tables. These two MOS include<br />

some 40,000 EM.<br />

.<br />

88<br />

_ ._ ..----.-.- -... .-. --__- --<br />

.<br />

.<br />

.


- .-_<br />

.!<br />

.<br />

.<br />

.<br />

__. _. . ._<br />

1<br />

_____.___ .-.-_.. . - . ..--- --.---. -. -- --- -.-<br />

. .<br />

--_ .-,,__ -____ -2 _._. _,_ . I ,-. . -.<br />

-._- --_a _-._____ _________, __-_- -..____._ .--a. .-._ . --.- -<br />

�<br />

.<br />

.<br />

MOS 94l.i, cook6 .<br />

Inrpection of the colum Gic’6 shown in Table 1 suggert that item6<br />

above a readability level of 9th grade fail to predict the criterion.<br />

The ric for 53 lteme with RI greater than 9th grade lo. .006, which ie<br />

not significantly different from zero. The ?i, for the 44 items with<br />

RI of 9 or lesr is .OSl. Thi6 Tic ir 6igdfiC6ntt,~ different from<br />

�����<br />

The difference between the two ‘Tic@6 is significant at the .Ol<br />

level.<br />

It is therefore concluded that the 53 item6 with RI greater than<br />

9 contribute little to the volidity of this MOS Evaluation Teet.<br />

The obtained validity of the test was .lO, Whatever validity the<br />

test ha8 appear6 to cane primarily from itUiI6 with lower level6 of<br />

readability.<br />

MOS 111.6, Light Weapon6 Infant-n<br />

Result6 for the LOO technical item6 are rhown at the left of Teblo<br />

If, and reault6 for the 25 supervisory item6 are ahown on the right.<br />

In the technical test,‘18 ittD6 with RI of 20 or higher have an<br />

a&rage validity of 0.014. The validity of the te6t, therefore,-which<br />

is .25, derive6 principally from item6 with a RI oelow 20. The ric<br />

for it-6 with RI below 20 is .065, which is significantly different<br />

from zero. The difference between the two riC’6 is significant at<br />

the .05 level.<br />

The supervieory tert can be divided st 14 itIm6 with RI up to 13,<br />

and 11 items with RI of 14 or higher. The Fi,‘s are respectively .086<br />

and .035. The former ie oignificantly different frcm zero and the<br />

latter i6 not. The difference between them ie t!ot rtati6tically significant.<br />

If the trend6 reflected in the rerults obtained to date are born<br />

out by additional MS 6emple6, it is believed that a strong relationship<br />

will have been demonstrated between the readability of teet items and<br />

item-criterion correlations. Step6 can then be taken to write test<br />

item6 at appropriate levels of readability. The result should be a<br />

significant increase in the power of items to predict job proficiency<br />

criteria.<br />

89<br />

.<br />

,*<br />

.<br />

.<br />

.<br />

/<br />

,.<br />

.<br />

I I


.=-<br />

. -’<br />

_<br />

,“ .<br />

.<br />

-<br />

. . .<br />

_- - . -.<br />

,<br />

-- .<br />

. . e<br />

. _ . ._~ -------.-..-<br />

- .A: .__ ._ ._<br />

_<br />

.<br />

L<br />

TABLE I<br />

MOS 94tl COOK N= 129 (3peer ratings) Apt. Area GT<br />

. .<br />

.<br />

Va I.= . IO<br />

24<br />

21<br />

Rt<br />

It3<br />

6<br />

RIla $ f z<br />

+tc .04 .06 .05 .05<br />

QC .05l<br />

tf 3.92<br />

p< .OOOl<br />

Omltted “M- 1337”<br />

beg* 8<br />

02 DO .02 DO -.02 -.06 .Ol<br />

.006 2 =2.50<br />

p< .OI<br />

.<br />

.<br />

-


s-t g<br />

CD L-9 "0<br />

aI<br />

- - -I -, 6-8 g<br />

4 v<br />

N a<br />

C!li! if 1: I<br />

1 ! p-4. 1 ! q 1 p-y<br />

! -1 -1 -1 ‘?<br />

I 61-W<br />

6 --<br />

Ci:!! :I.<br />

I I TI I<br />

12-02 0, =<br />

=j I--- = 12-02 is<br />

.<br />

16<br />

,<br />

.<br />

___ I.. ‘s<br />

.-...C -q.-~...v..-- --.-___ - __ _ . . ^ .- . ..__._.-_-- ---.---.------ _ _ _.^__<br />

_ & ._I. .,<br />

. . .._ _-. .-..<br />

.<br />

.<br />

_ .-- -- ___-.. . _.._ - .-- .--.<br />

. .


Performance Check Tests<br />

CLAUDE P. BRIDGES<br />

US Army Enlisted Evaluation Center<br />

Background<br />

For some occupational specialties, one way of making significant<br />

inroads into the variance in job proficiency currently unmeasured would<br />

be to measure more directly and effectively proficiencies involving motor<br />

sKills.<br />

In some occupational specialties, physical manipulation skills (nctor<br />

skills or the product5 thereof) are crucial and important determinators<br />

of significant portions of job proficieacy levels. The development of<br />

adaptations of written test techniques uhic!r will adequately measure such<br />

manual skills, ,at least indirectly, msy be possible ultimately in some of<br />

these jobs. The correlation between a man’s knowledge and his use of the<br />

knowledge on the job often tends to be quite high. However, there are<br />

several occupational specialties in which direct measures of such physical<br />

manipulation skills still will be necessary for good coverage of the job.<br />

In spite of the difficulties inherent in the world-wide use of performance<br />

tests, it has been possible, in the US Army Enlisted Hilitary<br />

Oc-:upational Specialties (HOS) in which such skills are most basic to<br />

differences in job profit-ency, to develop performance tests which can<br />

be administered world-wide under standard conditions and the results<br />

evaluated in an acceptably standard manner, However, such tests are quite<br />

expensive to develop, administer, and evaluate--often prohibitively so.<br />

As a result, it is quite important that every effort be made to explore<br />

other possibilities for directly evaluating motor skills and their<br />

physical products and to include such new measurement techniques among<br />

the evaluation instruments.<br />

This paper presents one such technique and the procedure for developing<br />

the required instrument. 0rl;inally I called it a Performance Check List.<br />

However , thie term is already used commonly for a significantly different<br />

type of instrument. In order to emphasize one of its two most important<br />

characteristic5 I now refer to it a5 a “Performance Check Test,” or in<br />

short a “PCT . ” On the basis of extensive experience in this measurement<br />

area, analysis of the pertinent reseqrch literature and logical considerations,<br />

the technique proposed herein can be expected to function effectively<br />

in the occupational specialties for vhich it is most appropriate, i.e. in<br />

the jobs which meet all the specified criteria.<br />

92<br />

�<br />

.<br />

.<br />

I


i’<br />

I<br />

.<br />

.<br />

When developed with ing;t,2uity, we may be surprised at some of the<br />

areas in which PCT’s would be appropriate. For example, Dr. Owen Waveless,<br />

an international specialist in linguistics, assures me that the development<br />

of an Instrument along these lines for interpreters should be ponsible.<br />

He said that such an instrument could be used effectively by an observer<br />

NOT skilled In a given language to evaluate an axaminee’a ~aneuaw’imJficiency.<br />

Even though it would NOT be a canpletely adequate aubstltute<br />

for more precise measurlas yielded by a good performance test, it would<br />

be much better than a traditional rating scale alone.<br />

. WHAT IS A PERFORIWXE CHECK TEST?<br />

- -<br />

.<br />

Y<br />

The proposed instrument is a true test developed In much the same wa,y<br />

aa a good written test, but the items are answered about--not by--the examinee<br />

and each item describes one specific mnnipulation, skill, or<br />

product that is marked by the observer, or supervisor, as being observed,<br />

or not -brerved in the examinee’s performance of the specified tasks of<br />

his job. It differs from the usual performance check list in that each<br />

item has been selected on the basis of experimental analysis (using<br />

standard item analysis techniques) frcxn a much larger list of specific<br />

proficiencies that first, can be readily and directly observed or<br />

inferred, second, can be so defined that any qualified observer would<br />

consistently state that the soldier observed either did or did not odequcll;ely<br />

perform the defined manipulation,and third, which distinguish between<br />

characteristics of various levels of proficiency. In some instances the ’<br />

manipulation described by the item might be a specific task. H,-wever,<br />

most items wou!d be limited to distinct individual skills or manipulation5<br />

involved in completing such tasks. For example, one item in a performanoe<br />

check test for a bandsman might be,“Consistently produces high C on his<br />

inetrument when called upon to do so.” A bandmaster should be able to<br />

answer this question about every one of his bandsmen, even without any<br />

special observation of them.<br />

Different items should describe consistently observable levels of<br />

skills possessed by examineee with different level5 of job proficiency,<br />

Some items should describe skills whfch are possessed only by the most<br />

proficient soldier in the MOS; Borne should cover skill8 which are<br />

porseoeed by the highly proficient but not by a man with average prof<br />

iciency; come should cover skills possessed by the average man but not<br />

poclstseed by those low in proficiency; a few should cover skill8 that a<br />

man low in proficiency can do but that a novice with some familiarity with<br />

the ffeld can not do at the level described by the item. Naturally,<br />

neither items covering things which practically everyone in the job can<br />

do, nor items which no one can do should be included for measurement purposes,<br />

.<br />

93<br />

---<br />

-<br />

,<br />

, . !<br />

1 .<br />

I<br />

i<br />

6<br />

,


,<br />

s<br />

.<br />

. ---_.<br />

. . .<br />

.<br />

.<br />

- -.--. .-. ._- .*__ ____-.._ -.-..-- -- --.. _.-<br />

“-. _-- ._-__.- .- ---<br />

Ideally, the items should NOT require the observer to evaluate the<br />

level at which the individual performs the skill covered by the item.<br />

In ‘other words, it should not be necessary to make qualitative judgment<br />

or inferences in deciding uhether or not a given exadntc performs the<br />

skill in the manner defined by the item.<br />

In btmary, the most important distinguishing characteristics of a<br />

Performance Check Test are that (1) the performances can readily be observed,<br />

(2) they can be consistently judged as being present to the degree described,<br />

or absent, and (3) they are selected by appropriate item analysis and<br />

validation techniques.<br />

The Performance Check Test might be considered as consisting of a<br />

special type of rating scales. Honever, the PCT is not a usual type of<br />

rating scale. It is a li.st of things which an observer can easily determine<br />

that a man can or cannot do. These should be weighted on the<br />

basis of experimental findings which will indicate each Item’s Importance<br />

as a component of the exsminee’s effectiveness cr job proficiency. The<br />

quantitative expression of the level of skill in the manual manipulative<br />

aspects of the job vould be obtained mensly by adding up the vei&tc<br />

assigned to the items checked for the individual evaluated on the<br />

Performance Check Test. A total score reliably and validly reflecting<br />

the exsminee’s skill is obtained. It covers discriminating motor aspects<br />

, of the job. When appropriately weighted and combined with the written<br />

test and other job,proficiency evaluations, a significant increase and<br />

in some jobs a considerable increase in validfty should result. Thus<br />

additional inroads into currently unmeasured variance in job proficiency<br />

should result.<br />

SlJHKARY OF DEVELOPMENTAL PROCEDURE<br />

1. ldenttfy the jobs for which current evaluation instruments are<br />

least valid in which motor skills or their product are important<br />

discriminators of d!.fferences in levels of job amstery, and in which<br />

the tasks involved most adequately meet the criteria, (See Annex 1).<br />

2. Analyze all of the pertinent information available or readily<br />

attainable and prepare a preliminary list of items to serve as an<br />

illustration for the subject-matter experts engaged in developing a long<br />

experimental list of potential PCT items. The “critical incident” approach<br />

developed by Dr. John C. Flanagan would be quite helpful. The “essay<br />

approach” proposed by Hr. Roberts, an USARES Supervisory Research Psychologi at,<br />

and his staff should be another helpful source of items. (See Annex 2 for<br />

their materials.)<br />

3. Convey these examples together with guidelines and appropriate<br />

accessory materials to the appropriate experts on the occupational<br />

specialty involved in the organizations at which such experimental items<br />

94<br />

.<br />

._


.<br />

.<br />

are to be developed. Discus6 with the subject-matter experts the posl;ible<br />

variation8 of the procedure 80, that they have clearly in mind what is<br />

deefred and how to provide it. (See Annex 3 for a list of criteria for<br />

m iteIII8.)<br />

4. Develop an experimental form of the PCT, coixsidering the "item<br />

criteria” lieted in Annex 3 and including all of the Items which are judged<br />

to be promising.<br />

5. Administer this experfmental form to a group of men working in<br />

the occupational specialty, a8 close as fe feasible to the date on which<br />

the specialty is being evaluated by the regular instruments. In the<br />

rample used, all levels of proficiency should be reprerented. At least<br />

two, and preferably three or more-observers of each men should apply the<br />

experimental PCT. At the 8ame time the be8t posrible externel criterion1<br />

data should be obtained,<br />

6. Analyze the item8 and the possible ways of scoring each against<br />

the total score on the experimental form of the PCT, against the written<br />

test, and against the external criterion. Obtain correlation8 also with<br />

all of the regular evaluation instruments. The consistency with which<br />

each ttw can be used will be analyzed. Techniques similar to thorc, used1<br />

fa refining biographical information blank8 and to regular item analyser<br />

should be applied. The desirability of assigning other than unitary<br />

weights, to the individu81 item8 In the check teat, should be determined<br />

empirically. The most valid and relipble scoring points for each item<br />

should lfkewise be determined empirically.<br />

7. Coneiderfng all the available data and the pertinent general<br />

measurement criteria, develop the revised form of the Performance Check<br />

Teat. If extensive modification of a eignlficant proportion of the experimental<br />

items has not been neceeeary, a useful estimate of the reliability<br />

of the total score on the revised Performance Check Test @CT)<br />

can be obtained at this time. The data also can be analyzed in relation<br />

to minimum acceptable proficiency 8nd promotion. Optimal Weight6 of all<br />

the evalucltion instruments naturally should be determined for the experimental<br />

tryout population snd appropriate personnel action8 determined<br />

for critical “cut score8.”<br />

Por jobs in which it ir pO8Sible to obtain an adequate number of PCT<br />

items, each of which can be consistently observed in a performance as<br />

meeting or not meeting the deecribed limits--e.g. for which the observer<br />

can reliably say,“This man can, (0. cannot) do this thing as specified,”<br />

and when this crucial statistical .-nalysis la made, the effectivenees<br />

of this Performance Check Test technique should be assured, The percentage<br />

95<br />

.


. _<br />

.<br />

‘. -7. . . .<br />

__ ‘. � ��<br />

���<br />

� �<br />

:<br />

�<br />

� �<br />

�<br />

�<br />

�<br />

�<br />

.<br />

__ . -.-_- _... . . . . ..- :<br />

_..-. -^-. -_ -. -... -<br />

of the veriancc in job proficiency mearured by the battery of evaluation<br />

inrtrumanf~ rhould be appreciably increased in any job in which motor<br />

� kille or their direct product are rignlficant detemlnerm of differencea<br />

in job/master, and thum in job proficiency.<br />

Reference<br />

Adkinr, Dorothy C., Primoff, E. S., McAdoo, H. L., Bridges, C. P., &<br />

Form, B. Conrtructioa and analysir of achievement teata.<br />

Warhlngton, D. C.: ;IS Go~nment Prfncng Officr, 1947, 211-265.<br />

96<br />

_. _ _-_ _.. -_ .- ---.- _-__. .._ --- --.- --<br />

. .<br />

*<br />

.<br />

: .<br />

. .<br />

.<br />

.<br />

,<br />

. .<br />

:<br />

* ‘% . -<br />

.<br />

.<br />

.<br />

.


---<br />

.<br />

.<br />

.<br />

.<br />

Aritm 1<br />

CRTTRRU FOR IDXNTIFYINC APPROPRIATE OCCUPATIONAL SPECIALTIES<br />

The general cha=actcristics of occupational apacialties in vhich<br />

performance check test6 might be expected to make a significant contrlbutfon<br />

tovard Increased validity are mmarizcd belov:<br />

a. Performance Check Tests should be considered for occupational<br />

specialties in which special motor skills or the direct product of such<br />

skills are critically involved. These rkille must NOT be comton to most<br />

men. Protzlnent among such occupational specialtie8, probably should be<br />

those having as their principal duty the operation, and perhaps in some<br />

instancea the maintenance of equipment.<br />

b. The motor okills ig the job must be consistently end readily<br />

obrervable to a. clearly definable level by an adequately informed evaluator.<br />

c. The rotor skill8 in the job must be crucial determinators of<br />

differences in level8 of proficiency, I.e., the physical skills must play<br />

an important role in discriminating betveen relative level8 of proficiency.<br />

Frequently MOST of the duties of such HOS vi11 entail special physical<br />

okills employed In the manipulation, use, and/or adjustment of toolfi,<br />

equipment, mechanism end/or meter:als. Typical are the equi-ent<br />

operator’e, .drivers, mechanic8 . machinfs t8, bandamen, fmne maintenance<br />

technicians, and 8ome automatLc data processing equipment technicians.


- --<br />

..a--<br />

\<br />

----+.<br />

-<br />

- i_<br />

7<br />

. . .<br />

- -7<br />

.)<br />

.<br />

-‘L. ..I<br />

yi-<br />

-+<br />

._ ..;<br />

. .._.<br />

.-<br />

‘..<br />

. . . +<br />

. ... ._<br />

- a.<br />

- s.<br />

,<br />

-7 ‘_..<br />

A<br />

.-<br />

,.<br />

--<br />

_. .-<br />

.-<br />

.-<<br />

.?-F-<br />

--CL<br />

_ .:-’<br />

- :<br />

.- .,<br />

_.<br />

.- _-.<br />

- _<br />

.<br />

“.<br />

:m.*<br />

,- <<br />

‘.<br />

$<br />

*<br />

.<br />

ANNEX2<br />

EXPIANATION<br />

Preeently, bandsman performance tests are being administered by<br />

having bandsmen tape record certain standard musical passages. The tape<br />

recordings are, then evaluated by specially selected and trained audition<br />

boards. This method of performance testing is time consuming, expencive,<br />

and pose8 many problem8 from an administration and scoring viewpoint.<br />

Any difference8 in scoring from one audition board to another decreases<br />

the reliability of the acore proportionally.<br />

The US Army Enlirted Evaluation Center ia conducting a study<br />

concerned with the development of a bandsman performance evaluation scale(s)<br />

which can be completed while the supervisor observes the bandsman performing<br />

hi8 duties fran day to day and/or by having the individual<br />

bandsman play special music passages or perform special duties.<br />

To construct bandsman performance evaluation scales, it is necessary<br />

to obtain information from as many sources as possible concerning the<br />

fmportant specific performance behaviors that are demonstrated by both<br />

good and poor bandsmen in the performance of their required duties. XndiVidual8<br />

holding a particular MOS are among the beot equipped to know<br />

what are the important types of behavior. One way of collecting this<br />

information from the individua’l bandsman ie to have him write a short<br />

essay about the individual he consider8 to be the best performer in his<br />

ttOS and to urite a short eosay about the individual he consider8 to be<br />

the poorest performer in his MOS.<br />

You are requested to write essays on the attached sheets in accordance<br />

with the brief directions given at the top of the respective sheets. One<br />

sheet la to describe the best performer you know; the other sheet is to<br />

describe the poorest performer you know.<br />

98<br />

. .<br />

_-. . _ . .<br />

__ __- _.-_ . -.__-- -.<br />

c<br />

,<br />

.<br />

-<br />

.<br />

.,‘!<br />

.<br />

.<br />

.


Page 1<br />

GR4ia --<br />

HOS --<br />

Length of<br />

Time XOS Held<br />

Total Years Hueic<br />

Training<br />

--<br />

Total Year8 Experience<br />

8a a Uuafcian --<br />

On thi8 page write a rhort essay describing the on-the-job performance<br />

behavior ,of the bent bandsman you know in your MOS. Emphasize the thing8<br />

he can do bttterxn mo8t in the HOS or can do that others crnnot. It is<br />

important that you write your array from your own viewpoint. D o n o t<br />

consult othero. to determine what you should write. The ‘combining7<br />

individual viewpoint8 will be much more v8luablt than having group<br />

opinion8 prtrtnttd a8 individual viewpoints. Write on the back of the<br />

page if nece88ary or use an cdditional oheet of paper.<br />

99


. .<br />

- _ ..‘.,. _.I_ -- ..- . ___-- - - - - .---TV-.---. -<br />

_ _______.-.... ?: _____.-____ __ ,-,. - ..,-.L4.. .--a I.----<br />

.<br />

: :<br />

. ---,<br />

. . . . - -- -- -<br />

..: ._<br />

.< ‘*<br />

. w<br />

.<br />

. . _i~<br />

.<br />

P&go 2<br />

‘.<br />

MOS<br />

Length of<br />

Time KOS Held<br />

On thir page write a rhort array describing the on-the-job performance<br />

behavior of the poorest Emphasixe bandsman you know in your MOS.<br />

thu thfngr he cannot do at all or as well aa the typical bandsman in the<br />

was. It is important that you write your era&y frw your own viewpoint.<br />

Do not conrult others to determine what you should write. The combining<br />

of Fndividual vlevpofntcl will be rrmch more valuable than having group<br />

opinionr pre’sentcd as individual vievpoints. Write on the back of the<br />

page if necerrary or use an addittonal sheet of paper.<br />

.<br />

.<br />

.<br />

.


.<br />

.<br />

ANNEX 3<br />

CRITERIA FCR PERFORMANCE CHECK TEST ITEHS<br />

1. Items should requfre as little judgment as to quality as possible.<br />

Ideally, each item would be evaluated as being “possessed” or “not possessed”<br />

by examinee.<br />

2. Item6 should emphasize licrftical’V motor skills that discriminate<br />

between various levels of performance.<br />

3. Readily observable differences should exist in the motor tasks,<br />

(or portion thereof, described by the item.<br />

4. The performance can be evaluated objectively in that it is<br />

directly measurable (i.e. can be evaluated as to percent accuracy of<br />

performance , count can be made of specific number of times something was<br />

done, is.performed at the defined level, etc., rather than having to<br />

evaluate it in a highly subjective manner where the standard6 will vary<br />

greatly from one observer to another.)<br />

5. Items selected for objective evaluation should be a representative<br />

trampling or cross section of the overall performance in order that the<br />

evaluation will not be one-sided or unfair to 6ome.<br />

6. Items of performance selected should be those that can be<br />

Ipresented in a uniform manner by different observers at different locations.<br />

7. Items should be those which when being used will have a high<br />

degree of agreement among observers.<br />

8. Items requiring a special performance of standard task may be used<br />

when required, but otherwise should be avoided to simplify completion of<br />

the PCT.<br />

9. Item6 which can be evaluated better and more efficiently by paper<br />

and pencil testing should not be included in this list. No item should<br />

have a lower correlation withtotal PCT score than with written test score.<br />

10. Items selected should be considered in terms of use for minimum<br />

qualification, proficiency, and pranotion score determination. This implies<br />

that items may need to be restricted to activities commonly per-<br />

IEormed by most or all.<br />

11. PCT items must conform with usual measurement principles and<br />

item and test criteria.<br />

.<br />

.<br />

,<br />

.<br />

:<br />

i


*. .<br />

_-<br />

*..<br />

--<br />

_-<br />

. . .<br />

:<br />

1.<br />

.-<br />

\ .<br />

L-<br />

:<br />

_ :a.-.<br />

:<br />

-_<br />

- .<br />

-_<br />

\’<br />

._<br />

-0.<br />

-.<br />

,.<br />

---.<br />

>.<br />

*\.<br />

-.<br />

T.<br />

,-- :<br />

&ntroduction<br />

Reliability of Checking Computer<br />

Produced StatfstfcS<br />

CASMER ‘S. WINIEWICZ<br />

US Naval Examining Center<br />

With the advent of computer processing, it became porrible to reduce<br />

large amounts of raw data into more easily handled and more meaningful<br />

forms. Previously, where estimates of the statistical characteristics<br />

of examination6 , such as Iawshe’s D or Flanagan’s r based on eithtr upper<br />

and lover 25 or 27% of a sample wtrt used, it became possible to awitch<br />

to the original point-biserial correlation concept bartd cm the total<br />

samplt. Whtre computer configurations are extended to include a tape<br />

system, it becomes possible to expand the samplt to include the entire<br />

group of examineeo taking a particular examination.<br />

Computer processing, when applied to item analysir, results in<br />

dtcrtated processing time and earlier availability of results. Itan<br />

correlations and difficulty indices based on the total sample result6<br />

in more precise measurement, while the expansion of the sample size to<br />

include the entire population results in even greater stability of<br />

measurement.<br />

Howtver, computer procesring has not been an L;unixtd blessing. With<br />

the onset of any new system, new problems are generated. One of these<br />

problems concerns the introduction of discrepancies into item analyoir<br />

results.<br />

The carue of these discrepancies can be divided into three general<br />

. areas � The first arta could be labeled Program Error. Although new<br />

computer programs are &toted on actual samples of the data they will<br />

process, there is a limit to the exhaustiveness of these ttst8. In<br />

areas of measurement where samples vary in size and characteristics,<br />

unforseen program difficulties may arise.<br />

The stcond arta might be called Machine Error. All machines, even<br />

those components with few or no moving parts, are subject to failure<br />

when used over a period of time. In the read and punch components of<br />

the computtr configuration, burned out circuits may occur, causing a<br />

failure to read or punch out particular card columns. Such discrepancies<br />

are generally consistent in afftcting a particular part of the itan<br />

a.nalyais processed before the failure is discovered. Ir the computtr,<br />

components cooling fans my fall causing over-heating which reeults in<br />

the variable .loea of data. This type of error is the most difficult<br />

to spot by visual inspection because it is variable.<br />

_._ . - --<br />

.<br />

. .<br />

.<br />

3<br />

102<br />

*_<br />

.<br />

-<br />

-t<br />

.


. .<br />

‘-<br />

-.-<br />

.<br />

.I :<br />

.<br />

-..<br />

(. .-<br />

,.’<br />

.‘;. . . .<br />

:<br />

.._<br />

. .<br />

../<br />

./’<br />

:<br />

.<br />

-.<br />

:’<br />

. .<br />

. ‘,\<br />

.’ ‘\<br />

.<br />

,<br />

.<br />

.<br />

.<br />

._<br />

������ ������ �������� ���������� ⌧ ���� ������ � �� �� � ��<br />

��<br />

- ~- . ---. .*._ -..-___ _,- .__^.. ____.<br />

.<br />

.<br />

���<br />

� �<br />

�� � ���� � �������� � � �������<br />

.<br />

.I..-.-..--.---. .._... _ __; _.., ._ ._<br />

c<br />

������� ��<br />

The third general area fall8 into the Human Error category. ~x~plea of<br />

thir type of error will vary from the use of the wrong cnsver key (eaey<br />

to spot vicually) to the inclusion in the ramplt of answer cards from<br />

another examination, or lack of coordination portafning to different<br />

elements of the program.<br />

In short, diacrcpancier do occur in machine-produced item analysis.<br />

and it is generally agreed that human inspection of item analysis rt6UltiB<br />

ir ncca88ary.<br />

Hethodr<br />

At this point it might be useful to conaider scme methods upon which<br />

vieual fnrpection of machine itear � naly8i8 re8ulte may be baaed.<br />

The firrrt method considered might be called the u8e of an outside<br />

criteria. One such outside criterion vould concern the Item difficulty<br />

index or item p-value. The p-value criterion 8tAte8 that a p-value may<br />

not be less than 0.00, nor greater than +l.OO. That In, not leaa than<br />

zero percent of the rample may an8wer an item correctly, nor more than<br />

100 percent. Another outside criterion would 8tate that the item-teat<br />

correlation or r-value (vhich 18 generalized a8 the discrimination index)<br />

may not be lese than -1.00 nor more than +l.OO without further invertigat<br />

ion. Another method of check:ng itan analyrir might be called the<br />

intra-item conristency method. If the item r and p values are bared on<br />

the total sample, and if the correct and incorrect item alternative8<br />

r and y ertimatar are baaed on a high and low group (upper and lower<br />

27% for fnrtrncc) then the correct alternative is being mearur.-d indcpendently<br />

on two overlapping groups to yield similar although not identical<br />

rerults. Empirical studier based on a large number of item8 will yeild<br />

probability table8 for variourr degree8 of diacrepancice betvecn there<br />

tvo ret8 of r and p valuer.<br />

A more involved method of checking item analysis might be termed<br />

the Item history method. This method is rertricted to pre-tested or<br />

control items, and COn8it3t8 of observing change8 in the p and r value8<br />

of the right and wrong alternative8 of an item from one examination<br />

period to another. Although the p-value of an item is canputed independently<br />

of other items, it8 value may change due to change8 in the examiaee<br />

population or varying degrees of item coatpromise or obsolescence. The<br />

r-value vould also be expected to have a history showing a degree of<br />

fluctuation which would be further affected by its dependence on the<br />

action of other itanu in the examination.<br />

Another method of checking item analysis data ir based on trends in<br />

the item results. When the p and r value8 of the examination items arc<br />

viewed am a whole, trendo may become apparent. For instance, if one or<br />

more clusters of negatfve r-values are obrerved wfthout cmpcnsatfng<br />

group8 of high r-values and the rav score standard deviation io adequate<br />

on the ballsi of experience or a rtatistical model, these trend8 may<br />

indicate error8 in the item analysis resultr.<br />

. .<br />

.<br />

1 0 3<br />

I<br />

� �<br />

� �<br />

f


c<br />

.”<br />

-. f<br />

. . . I...<br />

..’<br />

A varietioa of this method would consist of averaging the r-velucr<br />

� ppropriately end compering this average with the raw score standard<br />

deviation.<br />

These methods have two important chorattarirtice in couxon. Fir&,<br />

they lack prtciaioa. The Outoide Critaria, Intra-Itcln Conoiatency, Itan<br />

Hirtory, and Trend Method of checking iton anaiyeie irrvolve limits or<br />

ranges vithin which machine produced it- enalynie data CM vary for rea-<br />

OORO other than error. Although probability estimate6 may be determined<br />

for individual itgls variations within a “range” oc the basis of probrbilfty<br />

tables, the problem would remain as to tha range of probability estimatea<br />

which are apt to lndicata error in the iten enslyair data.<br />

The second camnon charocterietic of imporinnct concerns the expense<br />

involved in these methods. Checking the data, item by item, can be time<br />

consuming end require@ the time and effort of professional personnel.<br />

Extlnple of Interprogram Consistency<br />

A method of checking item analysis which may be called en interprogram<br />

consistency method has the advantages of both speed and prtclafon<br />

for computer application.<br />

This method a8aumta that there are tvo computer programa producing<br />

results on a given set of data. One program mill produce a raw score<br />

mean md standard deviation, and the other the item onalyeia data.<br />

The comparison of the result6 of the two programs 16 baro,d on the<br />

\ relationship of the sum of the item p-vaiues and the raw score mean, and<br />

the � ura of the ites reliabil!..y indicts end the raw &core etandord deviation.<br />

Two further arrumptionr concerning theaa rtatistical characterfrtica<br />

are that r;v ecorea ere computed on the barir of the number of correct<br />

� n8vera and ths;t the item discrimination index io a point-biatriai corrtlatloc<br />

(Gulliksen, 1958).<br />

me computetion of the sum of p and the sum of the reliability<br />

indlcer involves a third cmputer progrem which may be called the itemtent<br />

covariance progrem. The computation0 involvs summing the item pvaluer<br />

(Cp) to product a mean , end the une of the formula<br />

K<br />

L/T-G=<br />

1<br />

to product a rtandard devfation both derived from the itan e~IelyOi8<br />

(Gul’Llkrtn, 1958). The formula dm defines the index (&llikren,<br />

i958).<br />

-<br />

.<br />

. .<br />

104 -<br />

._ .- . . ---. ._~<br />

c<br />

, .<br />

.<br />

.<br />

.


. .<br />

.- . -<br />

.,<br />

.-<br />

2<br />

:.<br />

-W-.l.W.‘.. --c*--L- I_,__,__,_ __, __L I .- ..-.,_.-. “.. ..-. _ ,*. - .--. -<br />

.<br />

A comparison of the two Set8 or’ statbtics will reveal differences<br />

due to rounding error. The amount of this rounding error 18 basically<br />

controlled by the number of decimal placea compb:atfonr are curried out<br />

to before the rounding process is introduced. The raw acore mean and<br />

standard deviation will contain rounding errors 8~ Viii each itcm’p-.<br />

value and point-biserial correlation. The production and surming of<br />

reliability Fndictr vi11 result in varying mounts of rounding error<br />

in the derived mean and ottndard deviation. Consequently, the difference<br />

batvttn tha derived mean and standard deviation and the mean and standard<br />

* deviation based on the raw 8COitl vi11 vary froai ttlt to tert.<br />

One method that may be used to interpret the tlgniflconce of these<br />

difference8 conaisto of developing an empirical probability table of<br />

differences bared on data of known reliability. In its simplest form,<br />

thlr table would be produced by deriving means and standard deviations<br />

from a rample of one to two hundred eets of item analysis data, then<br />

comparing them vith mt@nS and standard deviations developed fran their<br />

respective raw score di&tributiona. The mean and standard deviation<br />

of the distribution of abroluta differences would form a rough probabi1it.y<br />

table which may be used In evaluating re8ult8 from future u&e of the<br />

interprogram method.<br />

If canputations art carried cut to the fifth decimal places before<br />

rounding, the difference between there #et6 of mtdns and etandard &viationo<br />

due to rounding error vi11 be minimal. Consequently, any difficulty<br />

that may occur in any of the three programs vi11 yicid rtsulte eubstantially<br />

outride those indicated in the rxnplrcal probability table.<br />

Under these conditions, the Jnttrprogram mtt’md can ba a prtcirt<br />

method of checking machine produced item analysis.<br />

The actual vorking time required, which includtr the running of<br />

thb* item-test covariance progrm to arrive at the final rtoultr, 18<br />

approximately one hour par thirty-two ratem, or thirty-tvo 150 quertion<br />

teats.<br />

The primary value of this method is to provide a precise and<br />

relatively quick check of sttm of itan analycllr data that will � tparatt<br />

out the nicety-five percent that is completely accurate.<br />

In thoaa instancea where the differences between the derived and<br />

raw #core � tsns and rtandard deviations art too great to be accounted<br />

for by rounding error, the difficulty may be in the raw score. the<br />

item analyria or derived rteult8, or finally an error In checking the<br />

derived results againrt the raw score results.<br />

Reference<br />

Cullikrtn, Ii., Theory of mental ttatu.<br />

- - -<br />

.<br />

105<br />

New York: Wiley, 1958.<br />

I<br />

..<br />

,


,b-.y<br />

.._<br />

HOS Evaluation Test Validation moiedurca<br />

RAYMIND 0. WALDKOETTER<br />

: US Army Eniirted Evaluation Center<br />

‘.\_<br />

‘-.._,<br />

\<br />

I’<br />

, -<br />

.<br />

,<br />

� � ����� ��<br />

The following points will be covered with the intent to give e<br />

condensed familiarization of this Center’o HOS Evaluation Past Procedures.<br />

The technical aspects of how it is done may be uxxe readily<br />

assimilated by checking vith the evaluation section and going to<br />

selected referencee.<br />

1. The emphasis on test validation has been reinforced by the use<br />

of the special rating of job perfomance as a criterion.<br />

2. The criterion is an appropriate rating sample of the job performance<br />

a8 experienced by peers under the guidance of the rating device.<br />

3. Test validation is concerned vith determining: first, just how<br />

the total evaluation test correlates vith the criterion; second, what<br />

makes up the valid portion and segments, and their individual and total<br />

correlations; and third, just how the outline for test developent can<br />

be used as a guide with the recommended nlnaber of item to increase<br />

validity.<br />

4. The validating procedure cmputations are coupleted with the<br />

multiple correlation between EZT, CER, and the criterion, wfth an additional<br />

validity coefficient given by the correlation between the weighted<br />

scoring formula and criterion.<br />

5. Validation activities will accelerate to give a hoped for<br />

iutprersive continuity in the qualitative and quantitative test control<br />

procedures.<br />

The achievement of MOS test validation has always been a basic task<br />

for USAEEC, but it has received a new impetus this paet March (1964) when<br />

the decision was reached to tise a newly drafted special rating of job<br />

perrformc:e as a criterion measure. Since, due to physical limitations,<br />

it is not immediately possible to validate all HOS, the HOS consequently<br />

oelected for validation were identified so that a maximun sampling of<br />

the personnel evaluated would be obtained during the prescribed test<br />

periods.<br />

A short treatment of the criterion is in order h,:ra. and possfbly<br />

a good word for peer ratings. An appropriate sampling of eM designated<br />

in specified HOS are rated by at least 3 co-workers and 1 eupervleor.<br />

Readability coefficients are estimated for ratings of each sample using<br />

a one-way analysis of variance (Wirier, 1964). The rateee must have been<br />

known by the raters at least one month, observed several times a week,<br />

�<br />

.<br />

��<br />

106<br />

_____. .._ _ . --__--I_-. ..-<br />

.<br />

, .<br />

�<br />

�����<br />

� �<br />

��<br />

.<br />

��


.<br />

.<br />

and performed .the duties 01 the Primary NOS. The ‘mean co-vorker rating<br />

given by 3 En for each of the men in the particular validation sample<br />

serves as the criterion. Since a sumnary of buddy ratings in military<br />

research by Hollander (1954) emphasized the relevant values of peer<br />

ratings, this information vae instrumental in the USAEEC research<br />

decision to apply the job performance criterion.<br />

An element of anxiety may surround the peer rating mthod in aom<br />

quarterr. It has been found wantfng because of ineffestfve application<br />

in most instances. For example, young recruits could not generally be<br />

expected to rate leadership in the military setting because they do not<br />

have adequate knowledge of the role and no degree of experience in making<br />

such a judgment. The fault lies not fn the racing process but tn the<br />

Ineffective preparation for fts use. The comparison most apt may be<br />

that of trying to ahoot a bull’s eye vfth a defective weapon. A stand wss<br />

made in our program to develop a rating form which would minimize rating<br />

format influences and insufficient knowledges of co-vorkers about each<br />

other. The rem1 t being, that peer ratings of a relatively structured<br />

style, are postulated with a sense of confidence toward obtatnfng a<br />

realistic estimate of job performance on an eleven-point scale.<br />

The administration of rating wa8 prefaced by special instructions<br />

to induce the raters’ acceptance of the task in a mre Informed and<br />

responsible may. A research psychologist conducted the rating session,<br />

while the teat control officer (‘X0) at esch installation we requested<br />

to schedule all of the available racers who were qualified to rate @I<br />

in the specified HOS. Croups of about 20 to 40 men were assigned to meet<br />

in suitable place8 for the rb;ing sessions, which were usually completed<br />

in about 20 minutes.<br />

Three phases of analysis compose the substance of the test valfdation<br />

procedure: (1) analysis of the total evaluation test; (2) analysis<br />

of the valid portion of the evaluation test; and (3) providtng rccomnencled<br />

numbers of item by evaluation test outline. These phases art organized<br />

to assist in recuring the desired evaluation test speclflcations through<br />

te8 t rev16 ion.<br />

Initially, in the first phase the relationships between item<br />

statistics and test statistics are thoroughly delineated. Results of<br />

aoalyeia in tabular form show the total test, technical test and Broad<br />

Subject-Hatter Areas (B!SMA’a) by number of items with the rtapective<br />

means, standard deviations, RR-20 reliability coefficients, validity<br />

coefficienta, beta weights, the multipltR,and coirected R after shrinkage.<br />

men follow in another table the correlation coefficients betveen each of<br />

the BStU’a, the BSHA’s and the criterion, and total evaluation test and<br />

criteriorr. In u third suxnary table the rerults reported for items<br />

include item p-vaiues for the total HOS population, p-values computed<br />

107<br />

:. .


.\<br />

.<br />

.<br />

.<br />

”<br />

for the validutioo e8mple, item standard deviations, item varibccee,<br />

item-test correlations, item-criterion correlations, indcxea of<br />

reliability, and indexes of validity. Mr. Wry and Hr. Shirkey h&ve<br />

presented in their reports, which treat the iten relationships, 8<br />

wider di6cour6c vith an8lyeis and implications in this ares using the<br />

needed examples for illustration.<br />

The second pha6e, concerning maximization of test validity by item<br />

alalection, ir appmached by tlimiruzttng item8 from the origin;ll evaiuation<br />

te6t which contributed nothing to validity. Correlation coefficient8<br />

appear in tabular form showing the relationehips between BSMA’s and each<br />

BSHA and the criterion vith the coefficient computed to indicate the<br />

validity of the revired evaluation teet.<br />

The third phase of the sequence Fn the telt validating procedure6<br />

occurs fn a m6ximlution of test validity wfth the optiml allocation6<br />

of items acccrding to the test outline. Techniques of multiple correlation<br />

uhfch determine optinval item ellocationa have been formulated<br />

by HOr8t (1949) and Taylor (195C). Theee technique8 reflect function6<br />

In relation to the correlations between B.SM’o, their reliabflitfcs, and<br />

validities. The correla t ions, reliabilities, and validitiec are systematically<br />

ctungcd. vhen the number of item per BSX4 are altered. By<br />

increseing the less reliable Bs;yA’s. validity is further enlarged, provided<br />

the valid vrriance is not measured alao by other BWA’r. After a clueter<br />

aMly6i6 (Br*Jchter. 1954) of KiNA’s the Horst technique is applied.<br />

(This procedure in applicable currently, but oay give way in deference to<br />

the Wherry - WIncr mathod for factoring large numbcr~ of items.) In<br />

tabular form a rwmary is given of the cluster analyeie showing each<br />

ciuster, optlpal nrrmber of item6 per cluster, and the BWA’s per cluster<br />

with the valid number of items In each BRIM, the proportlon of BStIh<br />

ltem6 in each cluster, and the optiraal number of iteuis per BSHA.<br />

From a ruemary reviev of the validating procedure8 used vfth the<br />

te6t analysis, a brief description of the overall validity approach fa<br />

desirable at this point. Since urch, some six MOS terts which eVAlUStU<br />

mboot 17% of EM under the Army EES, have received validity analysis and<br />

evaluation based upon the criterion of job performsme racingr.<br />

The ssmples used rnnge from 30 to 129 and were checked to assure<br />

reprerentative groupr. The evaluation test validity coefficients<br />

ranged from .lO to .52. The CER had validity coefficients rangihg from<br />

.20 t o .51. The multiple correlation coefficient6 betveen evaluation<br />

test, CER, and the criterion ranged from .36 to .S5sdemostrating a<br />

rlightly better prediction of validity from the BT and CER combined.<br />

The corrected multiple correlation coefficients after ehrinkqe range<br />

from .34 to .51.<br />

. . . .<br />

108<br />

- .I<br />

.<br />

.


’<br />

.<br />

I<br />

,<br />

._ ..--“.,.-.“..~...;..-. . . . . _ _. .<br />

.<br />

bt ft be remarked here that although some validity coefficier.ts<br />

.have not reached entirely sanctioned level8 of validity wet are<br />

at.taining significant levels of validity. With the identification of<br />

weak points in validity we can utilize greeter control thereby beginning<br />

the necessary corrective test developtint procedures. For example,<br />

the validity coefficient for the HOS 941.1 Waluatfon Test (Cook) wa6<br />

.l.O, vhich should be-susceptible to improventent by simply eelecting<br />

‘the proven valid items and adjusting the reading level of the test to<br />

that more usually experienced by cook6. Mom a mmparisoa of validity<br />

coefficients I would like to hazard a conjecture that the validity<br />

coefficients tend to be higher for the more technical jobs of HOS<br />

t!mn for the unskilled or perhaps motor-skilled jobs.<br />

Another combined validity coefficient fa given by the correlatfon<br />

between the rav composite score or scoring fomula and the criterfon.<br />

These coefficients for the six MOS ranged from .22 to .49.<br />

To realize that validation procedures fn &ma of man-hours, data<br />

puocessing, and interpretation presupposes a high degree of technical<br />

organization, requires very little analytical skill. But to utilize<br />

these procedures and insert the mdfficatfons needed to improve validatfon<br />

procedure6 demands that an organ?zatfon reach a noticeable stage<br />

of maturity. The maluJtlon and Ar~lysis Branch of USAEEC, nov embarked<br />

upon thts stage of maturity, r:il continue to improve and increase ft6<br />

vrlidation efforts fn 6a63p?.it!43 all area6 of the HOS 6trUCtUre.<br />

109<br />

, C<br />

.!<br />

i


References<br />

Pruchter B. Cluster enelysir. Introduction to fnctcr knalyaic.<br />

New York: D. Van Nostrand C-an>, Inc.,TF54, 12-17.<br />

Hollander, E. P. Buddy ratings: <strong>Military</strong> reosarch and induatrial<br />

implications. Personnel Paychol, 1954, 7, 385-393.<br />

Horst, P. Determination of optimal teat length to reurizaita the<br />

multiple correlation. Pepchometri&, l>G9, 14, 79-88.<br />

Taylor, C. W. Maximizing predictive efffc.ency for a fixed total<br />

testing tima. Psychometrfka. 1950, 2, 391-406.<br />

Wherry, R. J., and Wirier, B. J. A methqd for factoring large numbers<br />

of iteuxu. Paychometrika, 1953, 18, 361-179.<br />

Wirier, B. J. Single-factor experiment8 having repeated measures on the<br />

saxa8 elements. Statistical principles in experimental deeign.<br />

New York: McGraw-Hill, 1962, 105-139.<br />

I<br />

. s<br />

.<br />

110


.<br />

.<br />

.<br />

-<br />

. .<br />

. . -.. . ^. ‘. L ,... . . . . . j . .<br />

Background<br />

Approaches to Improved Measurement:<br />

Research in Progress<br />

. .<br />

CASLHER S. WINIF~ICZ<br />

US .Nsval Examining Center<br />

The research projects abstracted herein are primarily projects<br />

th8t the US Naval Examining Center has concentrated their efforts on<br />

sfnce the last <strong>Military</strong> <strong>Testing</strong> <strong>Association</strong> meetfng, held in 1963 at<br />

Groton, Connecticut. ThcSe.prOjeCtS should not be confused with the<br />

regular semiannual evaluations that are conducted on all examinations<br />

and their respective populatlon8. To eumarize a few of these autometic<br />

evaluations; each examination is anslyzed fn terms of it8 adherence to<br />

the criterion standard of construction; professional and military sections<br />

are analyzed independently and also theft interactions a8 major components<br />

0” the total instrument; intercorrelations between all minor aubsections<br />

on the professional pgrt, as well as vith the total composite;<br />

item atUllySi8; raw and standard score conversions along with their<br />

respective statfetics and graphic representation of each population a8<br />

well a8 the parameters; analysis of the final multiple components; check<br />

for compromise and collwion; and finally all the various activity and<br />

bureau reports that are required to adequsltely summarize thfe information.<br />

A susxaary listing of the following research projects fndir-pte those<br />

unique areas that h8ve been investigated to provide-additional technical<br />

Information TV improve and support the Naval Advancement System.<br />

Pro-Pay Survey<br />

A survey was conducted on the prevalent attitude toward a variable<br />

reenl!.stment bonus (guaranteed minimum bonus with a variable sum based<br />

on rating criticality) vice a proficiency pay program. Ten thousand<br />

canfdates were sampled from a cross-section of 231 different Naval Activitfee.<br />

The major extrapolation from the data tend to support the premise<br />

that,in general,the group would prefer a variable bonus system in lieu of the<br />

present pro-pay program. This I.8 based very briefly on the fact that<br />

although only 25% of the total group are currently receiving pro-pay in<br />

one form or another, they only favor pro-pay In a 60 to 30 ratio. However,<br />

the larger portion of the sample, the remaining 75% that represents about<br />

6000 people in this sutiey are in favor of the bonus over pro-pay in a 2<br />

to 1 majority.<br />

Automatic Sxamination Requisitioning<br />

The examination answer card8 have Jeen revised to Incorporate the<br />

collection of two variable factors of performance evaluation and awards.<br />

Basic battery scores ‘Ire constant, and length of service and time in rate<br />

*<br />

111<br />

I<br />

..- . - -_. ..-<br />

; -! :<br />

i i


.<br />

,<br />

,<br />

. -<br />

c<br />

increase by a constant for each examination seriee. A propoaed tape<br />

configuration s-1ded to the present computer would allow for the elimination<br />

of certain administrative responsibil.ities prior to the administration<br />

of each examination (elimination of !UV,PZRS 524 which contains<br />

biographical information on each candidate). A raster tape file cotttinuoualy<br />

updat,ed by NEC in conjunction with the collection of certain<br />

infonnntion directly from the answer card would eventually eliminate<br />

the necessity for ordering examinaticns and they could be sent out<br />

automatically by name vhen each candidate becarocs eligible, as indCcatRd<br />

by the file maintained by NEG.<br />

Actuarial Longevity Prediction Study<br />

Since the literature only covers conventional statistical approaches<br />

to prediction, such as regression analysis or multiple regression, and in<br />

some cases the application of the Poisaon distribution or the negative<br />

bi-nomial, raw data was acquired to experimentally determine the fcaafbility<br />

of utilizing the actuarial approach in predicting longevity. F31vfdence<br />

to data indicates that this approach will eventually be ctilized in<br />

certain areaa of prediction as a routine technique, To fully appreciate<br />

the success attained by this approach, addifional comparison studies<br />

between various predictive techniques will be conducted aa to the eventual<br />

superiority of one in terms of minimizing error, fewer assumptions, time<br />

clement, and the final extrapolations from the data that are poaeible.<br />

Class Scheduling Project<br />

A atudent, teacher, and class scheduling project was cmglcted by<br />

hand baaed on raw data (N=2000) for a single school, The application<br />

of resulta will eventually be duplicated by a computer program and the<br />

two aete of results will be analyzed for comparibility. The project<br />

will then be extended for application to the CNARESTRA problem (N-35,000)<br />

which will Involve a series of schools located geographically in various<br />

parts of the country and eventually encompass the mobility of atudenta<br />

from one area to another. Complete coordination will be poacible between<br />

the transportation, teacher, student, school, classroom, time sequence,<br />

and course information factors, all will be in one co;aplete computer<br />

program.<br />

Differential Weights of Final Multiple<br />

The final multiple acore is composed of five individual factor<br />

scores which are a-d on the basis of a simple weigLting formula.<br />

Each factor has a designated ideal contribution to the overall variance,<br />

but the actual or real factor contribution IZWIY and does obviously vary<br />

from this standard. A study was implemented to determine the empirical<br />

interaction of the five factors making up rhe total composite of the<br />

multiple variance of passing candidates in selected critfcal ratea frclm<br />

previous examination series. The results indicated that the factor<br />

112<br />

.


.<br />

.<br />

.<br />

contributions to the fir&l mulctple variance differ from the apparant<br />

atandard by rate and factor. @n the basis of pay grade medfans, differences<br />

from the standard were etatistfcally significaut for all factor8<br />

at pay gradeo E-4 and E-5 and not significant at pay grade E-6.<br />

Pre-<strong>Testing</strong> for Controlled ~minations<br />

The Navy’8 regular advancement examination is compoeed of 150<br />

4 option, multiple choice questiors. In the construction of theae<br />

instruments only 50% of the questions have item statistics and are<br />

utilized to control the standard of the examination agalnat the deoired<br />

criterion, The remainfw percentage of question8 are new and generally<br />

cover the entire spectrum of difficulty. Closer adherrnce.to overall<br />

a tandards can be acquired by simple pre-testing of the new questions<br />

on selected populations. This in eesence will provide for a completely<br />

controlled examination from the standpoint of item difficulty and discrimination.<br />

Reliability Computer Check Study<br />

i<br />

7.<br />

A detailed approach to one aspect of reliability checking for<br />

computer operation8 ie given in another part of this surmary. The<br />

paper wa8 presentad independently a8 prt of the Theoretical Seminar.<br />

Four Cycle Bcaminin~ Periods<br />

The regular advancement examfnattona are pr?marily discriminating<br />

and administered Navy-Wide each February and August. The regular input<br />

at pay grade E-4 cannot be maintained in 22 critical rates, vithout<br />

disturbing the balance of the total system. The possibility of examining<br />

these critical perrronnel duriw Hay and November with a qualifying type<br />

oli examination ha8 been investigated and proposed a8 a poseible solution.<br />

Simplified Tri-Serial Correlation<br />

A byproduct of a latger project yielded a simplified triserial<br />

correlation, result8 of which were presented at the national 1964 APA<br />

conference. J88pCn’8 original formula for the triserial r ie rather<br />

unwieldy and therefore has been passed by for 8Ome easier lees appropriate<br />

correlstion technique. Jaspen’s formula equals:<br />

‘tri p<br />

ZaYa + (Zb - za) Y b - ZbYc<br />

Za* + (Zb - Za)* +<br />

=r 8 b<br />

C<br />

113<br />

.<br />

.-<br />

.


,. . . ;<br />

./:.<br />

. .<br />

--.. ----.<br />

,.<br />

,.‘.<br />

5/’<br />

,./’ .<br />

\‘<br />

.\.,‘.<br />

2.<br />

-<br />

. _. ., _ .<br />

The formula derived which gives identical resuAts with Jaspen’e<br />

formula equals:<br />

I‘trf u<br />

CYa + (Na - Nu) (Yt) - L”y-2<br />

Ntuy (2, + 21)<br />

Weighting E Two Subteet composite<br />

Another byproduct and also presented at the APA, was the tteighting<br />

of individual subtests in I composite, where they become more critical QBL<br />

the total number of tests in the composite dscreaaes. The importance of<br />

controlling the weighting of individual eubtesta reaches a naxfmxu When<br />

there are only two subteeto in the composite. A method of controlling the<br />

contribution of each subtest in a two teat composite ia given by the<br />

following formula:<br />

Wa + rab<br />

Za<br />

’ + ‘ebWa p<br />

f-<br />

1 - za<br />

Open End Item Distractor Development<br />

Pay grade E-8 and g-9 have essentially an examination tbAt is part<br />

aptitude and part achievement. Some of the subteste contain leadership’<br />

and situational problem questions’which are extremely dfffic*tlt to<br />

develop. A technique of using an open end question to collect responses<br />

was utilized. ‘The various responses are then tabulated and arerbad into<br />

aimllar groupi. ge by a frequency count. The three most attractive incorrect<br />

rsplies are then merged with the correct anewer to form an ites along<br />

with the appropriate stem.<br />

Longitudinal QualFty Control<br />

A quality control study is routinely conducted after each advancement<br />

cycle in order to determine adherence to etandarda. It has the<br />

advantage of presenting current Fnformstion and aces as a warning system<br />

relative to overall examination quality. A lolyitudinal seudy based on<br />

previous quality control studies is now in progress and its effect will<br />

be cumulative in nature and will form the baeia to predict the desired<br />

atatfstfcal characteristics of future examinations,<br />

Effect of Automatic Advancement<br />

Various programs at the minimum petty officer level (pay grade E-4)<br />

were introduced to act as a stimulus for reenlistment in the &IVY. As<br />

a rettwlt of euch programs as STAR, SCORE, and CLASS A school, atitoPratic<br />

. .<br />

.<br />

.<br />

.


._<br />

_ --r<br />

-.<br />

:’ ,--_<br />

.I’-<br />

..__<br />

..- _<br />

_.<br />

.I. :<br />

‘\ ‘,.<br />

,<br />

.<br />

.<br />

--<br />

.<br />

-. .I<br />

. . .~ . .<br />

promotion to pay grade E-4 is possible without the necessity of taking<br />

an examination. Since the inception of these programs which are a<br />

recent innwat ion, a certain normal quantity of the examination population<br />

is missing. This mean8 inessence taking the quality off the top and<br />

having the remainder take examfnations. Tine standard at E-4 will obviously<br />

change and also when this scgt?ent of the population ia added back into the<br />

competitive examination popclation at the next higher level at E-5, certain<br />

ramifications will take place. This problem is currently being fnvestigated.<br />

Forced Sampling Technique<br />

The forced sampling technfque is a form of stratified sampling.<br />

It controls a sample by reducing biases caused by chance factors in<br />

random sampling. The procedure is to rank the total number of candidates<br />

from lw raw score to high raw score. The next procedure is to<br />

pull the sample in such a way that all raw scores are represented as<br />

equally as possible in the sample. The formula used to choose the<br />

individual samples was: 4Nt/N, + Nt/N, + Nt/Ns...Nt equals the total<br />

population, and Ns equals the sample size vhich is desired. The forced<br />

sampling procedure has been found to be extremely accurate in representing<br />

the population median mean, and standard deviation even with<br />

sample sites of less than 50.<br />

Population Analysis Instead of Random Sampling<br />

The sample sizes of the occupational rates are based on the limits<br />

of a confidence interval of a population. The confidence in the limits<br />

of a mean for a given parameter is the fiduciary probability. The<br />

fiduciary probability is better than .95 that the true mean 1i:e in the<br />

interval n + 2.00 S. E.,, and .OS that it falls outside the 1im:ts. A<br />

standard error of one raw score was the average standard error of our<br />

samples. The effects on item analysis ard reliability of the examination<br />

is noticeable, although by increasing the number of candidates used in<br />

the studies tne vatLance also increases slfghtly. However, the use of<br />

the total population in the studies provides more stability in all item<br />

statistics.<br />

Validity<br />

Concentrated effort has been placed on obtaining indirect measures<br />

of validity through analysis of class A school graduates versus non<br />

school graduates, frequency breakdowns of various elements of the population<br />

and conducting teats of significance on the differences, Conventional<br />

validity through supervisory ratings and peer ratings has<br />

yielded validity coefficients that average around .35.<br />

115<br />

.<br />

--- ‘-<br />

.


Approaches to Improved Measurement:<br />

Research in Progress<br />

RCMLD K, GOODNIGHT (<br />

US Army Enlfated Evaluation Center<br />

Throughout the confer&e we have heard and discussed numerous topic6<br />

epanning the realm of test immrovement and development. In every cane<br />

each method used, the reaulto obtained, and the utilization of these<br />

re6ults are all bared on reresrch. Reeearch is the guideline to BUECC~U<br />

in almoet any endeavor. Tnncrefore, I am plcaeed to present to you several<br />

of the more important reeearch projects which have been completed, as wel.1<br />

68 nome currently in progrerit, at the US Army Enlirted Evaluation Center.<br />

The first project, completed oeveral years ago, ir A Comparison of<br />

Seven Methods of Computing Total Test Reliability fromma Single Terr<br />

Adminietration. Ttdt reliability (that is--a determination of teat conriatency)<br />

is evaluated by Lnvtstigating tht relationship of the praiictor<br />

to itrtlf, or in other words, it is-the relationship of the ranking of<br />

scores on one administration with the ranking on a aubsequant adminlstratfon.<br />

The rsliabilfty of a test can generally be estimated adequately<br />

fran only one administration of a test.<br />

For tach MOS Evaluation Teat a reliability coefficient is computed<br />

via the Kuder-ttfchardoori Formula 20. However, it ~68 imperative tc know<br />

if thie wa6 the most accuratt indication of the consirtancy with which<br />

the tert was measuring job proficiency. Thtrefore, this project was<br />

designed to determine which of the odd-even, K-R 20 on the total group,<br />

K-R 20 on 54% of the group, Hoyt’s Ana1ysi.e of VerCanct. Horat Maximum.<br />

Horet Corrected, and Cleman’r Maximum reliabili*.v mearurement methods wan<br />

super for. The result8 showed the K-R 20 relfability sethod on the total<br />

group io the meet appropriate for the MOS Evaluation Te6ta, thus supporting<br />

It8 uragt by thL Enlisted Evaluation Center (EEC).<br />

Thi8 study was completed in February 1963, and prcaently another project,<br />

A Canpariron of Fivt Hethods of Computing Total Ttrt Reliability frus<br />

a Single Test Administration, ir under way. This 18 a replication of the<br />

earlier study with some minor variation6 tc further verify EECI currant<br />

procedure8 in tt6t reliability estimation. Also, another reason for con=<br />

ducting this study lies in the proposed technical rtcommendationr which weft<br />

presented at the 1964 American Psychological <strong>Association</strong> meeting&; their<br />

rscamntndationr indicate that the K-R 21 ia more appropriate to use then the<br />

K-R 20 under situations such as those at EEC. Therefore, in this rtudy five<br />

methods of reliability eatimation-- K-R 20, K-R 21, odd-even, Hoyt’8 Analysis<br />

of Variance, and Uorst’o Maximum--are beir.g compared to determine which is<br />

the better method of mearuring te6t cons,etency in view of APA’s rtccmmendations<br />

and to rtudy the effect8 of aample sire and MOS rkill level on the<br />

varioue reliability coefficlentr. Results are not available thus far.<br />

116<br />

-_ ._.....- .--- .---. --.<br />

c<br />

_I__...


,’<br />

‘.<br />

,-- -<br />

: ’<br />

‘.b<br />

:’<br />

_. _<br />

:<br />

:<br />

I ,<br />

_.<br />

. .<br />

f<br />

-.. \<br />

‘\<br />

-.<br />

‘.<br />

Another area of reoearch is the longitudinal studies on the cmmander’s<br />

Evaluation <strong>Report</strong> (CER). The CER is the officfal rating form<br />

which constitutes one of the cmpone’nte of the Enlfated t’valuation<br />

System. The primary purpose of the CER is to provide an sseessraent of<br />

the soldier’s job performance and potentinl-for advancement by his imdiate<br />

and secondary supervisors. Although ratings, per se, ‘are not perfect they<br />

are the best available evaluation method when no objective masure can be<br />

obtained; therefore, the enlisted Evaluation Center is quite interested<br />

in the perfomance and improvement of the CER.<br />

no studies presently beinS conducted to facilitate the continual<br />

improvement of the CER are A Factor Analysis of CER Scales, <strong>Technical</strong><br />

Subtest. and Suoervieorv Subtest. and<br />

I A CccJlparison of a Graphic Raa<br />

Method, A Normative Paired-Compartson Rating Method, and An IpeatFve<br />

Paired-Comparison Ratfng Method via the Multitraiz-Multimethod Hatrix.<br />

The factor analytic study was to determine the degree of c-n-factor<br />

variance of each of the twelve CER Pcalee and the technical and supervisory<br />

subteata ofanMOS Evaluat:on Test. prom the results of this<br />

research will come information necessary for proper revision of the CER<br />

to eliminate rhe measurement overlap or c-n-factor variance, thus<br />

improving the evaluative ability of the zating. Seven orthogonal factors<br />

emerged from the analysis, however, tbey a:1 have not been named. One of<br />

more important results thus far, however, was the high loadfng of the<br />

technical subtest on the initiative Factor.<br />

The second rating study, presently in the planning stages, ie A<br />

Comparison of a Graphic Rating Method, A Normative Paired-Cozparieon-<br />

Rating Method, and an Ipaative Paired-Comparison Rating Method via the<br />

Multitrsit-Hultim+thod rWtrix. This research endeavor was dec Lgned to<br />

statistically letemine whether the graphic rating method as used in CER<br />

ratings is adequately serving its purpose by providing valid and accurate<br />

information to the Enlisted Evaluation Center, or, if either the normative<br />

paired-canparison 4th certainty judgments method or the ipsative pairedcomparison<br />

with certainty judgments method would be superior and provide<br />

more accurate and valid results. Also, since the normative and ipaative<br />

data are aupplemntary to each other, it is possible to statistically<br />

combine the data from these two methods, thus yielding a fourth rating<br />

method for the analysis. ft is felt that this information will be<br />

valuable in assessing the adequacy of the CER ratfng method in comparison<br />

to the other rating methods. No data have been collected yet,<br />

One very important previuus study was The Effect of Rater-Ratee<br />

Acquaintance Period on CER Ratings. This research project was designed<br />

to determine ii a specified minimum period of acquaintance between the<br />

rater and ratee was neceesary for satisfactory and reliable CER ratings.<br />

The results obtained on various NOS skill levels indicated that a minimum<br />

period of two months (60 days) acquaintance was esaentfal for proper<br />

evaluation. This time period is now mandatory in the Department of the<br />

Amy for all CER ratings, although some leniency in this requirement is<br />

alloued.<br />

117<br />

-------<br />

.<br />

‘I<br />

;;


.\- .<br />

‘.<br />

,-.<br />

/<br />

I<br />

\ ,<br />

_. /’<br />

,<br />

/<br />

/’<br />

. .<br />

,_ -y...---<br />

. . --<br />

I<br />

.<br />

.<br />

.<br />

FoeoiL\‘ly the most important area of research at the Ealfated t-kaluation<br />

Center are, the validation ctudiee, of the Cczzindtr’i: ZvalusrLon<br />

,Repzt and the KG Evaluation Testa which ere conducted routinely.<br />

Formal val.idiry rcpmtlng is quite comprehanui:te and tncltrdes the<br />

validatioil of both the CER aitd the EvaloAt.fon Teat as *:ell as the interrelationship&<br />

between these instrrantc. Dee t o thrj lergthy processin,%<br />

t&e required for the forwl validity report, a prclimin~ry reporting<br />

procedure io employed to provide maxiliium data for te3t revicion.<br />

A preltmtnary validity report is to prwtda guidcltaea in test<br />

revision based on pertinent statistical deta. TF;fo vaiiGi.ty data<br />

cover&z’ the entire evalwtion test outlin 1) and individual test item<br />

are provided prior to test revision to allw for their UB~ in the<br />

decision-making process of test development. Emphasis is now placed<br />

on the WIfi and HOW of test vslidity.<br />

Statistical bases are provided to allow control in :c?~~t revhim<br />

over the evaluaticq test characteristics of wan, ste.ndard deviati.on,<br />

reliability, and v.llidity. The test characteristic of valfdity is<br />

given specfal atta.;tLon, and is considered at both the item and test<br />

outline levels. The statistical rationale and data necessary for thinpurpose<br />

are provided to enable their use In test develcpment procedures.<br />

Tha users of the information SGI data provided in the validation<br />

report should: (1) have a thorough working knout,cdye of the interrelttionshfpa<br />

between item and te6t stat!.st:cs to enable zzre control<br />

in test revision Over evaluation test means, atar.darJ devls.tkona,<br />

rcliabflfties, and validitlrs; (2) insofar a8 is ~racr.fcable, include<br />

items of substantial vaiic:ity in revised tests; (3) study itema of etibstanti.al<br />

valfdfty to deten?fne the particuler types of items which tend<br />

to be most valid far a given Rvaluation test; (4) ff ouClina revision<br />

is deeut?d appropriate, make such revisions in view of known intarrelatio.%hips<br />

between Eroad Subject-patter Areas; and (5) reconcile item<br />

requests vith both optin item allocations by Broad Subject-t!nttor<br />

Areas and considerations of n practical nature. In this way, mxtnwn<br />

control of the test results can be attefned.<br />

A Corcparative Study of a Short Form and a Lorq Porn of Performance<br />

- -<br />

Dictation Test for Legal Clerk or COure <strong>Report</strong>er - - te 8 research project<br />

presently being conducted at the Enlisted Evaluation Cuntar.<br />

his project was designed to detennfne whether a short form of the<br />

Dfctation Perfornnnce Teat may be relLable enough to use in place of the<br />

longer form, thus savfng administration and scoring time 88 vell as<br />

providing a more maruzgeable measurement of the selected examineen. fie<br />

short form is composed of preeelected sections of the dictation test<br />

crmaprising 4GX of the total test. In a preliminary aralyeis, a Peareon<br />

r correlation cofficiene of .95 was obtained bemeen the long form and<br />

the exparitintal short revfsion. It would appear further examination<br />

of this relationship wfll prove fruitful, and the recommendation for<br />

the shorter test can be anticipated.<br />

118<br />

-- .- , -. . ..____. _____-... - __ .- _..-- ___---- ___^____ l_--


.1 , .I . . ., .;., ., _-<br />

.<br />

-.< “. ,:. ,L ,._.,._. _<br />

Another group of continual research studies lies in the Inv’eetf-<br />

Bation of Possible Test Ccmpromiee. One study i.n this area vns the<br />

davelopwnt of investigation procedures.<br />

An MOS !Zvslustion Test would be comprolaiaed if one or more copltes<br />

, of the test booklet came into.tho possession of an unauthorized enlisted<br />

_ nun, and the Information vats used by him before or during the admints-<br />

/ tration of the test, An HOS Evaluation Test is subject to possible<br />

compromise if it is lost or is unaccounted for prior to the ccmpletion<br />

of an “02 evaluation p%riod. . .<br />

: . .<br />

. .<br />

., _,‘.<br />

.<br />

.<br />

In this study three methods for investigating possible compromise<br />

in a large military program in which tests are administered once a year<br />

were developed and experimentally tested. Compromise may be checked by<br />

(1) comparison of test scores of individuals or possible compromise groups<br />

with population parameters; (2) standardization and analysis of test<br />

scores wer two test periods; and (3) regression analysis of test scores<br />

over two test periods. prom the resulting statistical analyses the<br />

limitatfors and advantages of each method vere shovn, as well a8 a<br />

rationale for the interpretation of results and the formulation of<br />

subsequent administrative decisions and recolmlendations, Presently,<br />

any one of these metho,ds may be used depending on the circumstances.<br />

Two more recently carsrenced studies are A Comparison of Six Me,hoda<br />

of Item-Test Correlations and k Comparison E v a of%ur l u a t i o n Proceduref!<br />

for Pieasuring ParforuGG Efficiency. The first fesearch project was<br />

based on Cuilford’s (1950) study in which he compared the biaerfal r, the<br />

point-bieer?al r, the ordinary tetrachoric r, the Flanagan tetrachoric r,<br />

and two applications of the Phi Coefficient methc-;s of itein-tese correlation.<br />

~%a point-biseripl r proved to be the superior method. Therefore,<br />

thfe project was designed to replicate and extend Gutlford’e study<br />

since the point-bisarial correlation is sued in item analysis at EbC.<br />

This project will either substuntiate Guilford’s findings and give<br />

further proof of the value of the point-bisarial method, or it will<br />

reviee his results and indicate possibly another more accurate index of<br />

Item-teat correlation. ho data have been collected as yet.<br />

The second study noted above can be classified as a criterion study.<br />

Sufficient procedures are used by EEC for increasing and maintaining the<br />

reliability and content validity of the HOS Evaluation Tests. However,<br />

a number of problems are encountered in the establishment of concurrent<br />

or predfctive val:ldfty since an adequate criterion is necessary. The EEC<br />

has developed a rating scale which is presently being used for getting<br />

criterion data via co-worker or peer ratings. However, more satfsfactory<br />

criterion data msy be accessible by using other measurement procedures.<br />

The purpose of this project is then to compare four methods of<br />

measuring perfor=mancaefficiency to learn whether a superior rating<br />

119<br />

,


.L<br />

.: -_ _.<br />

.\<br />

.‘? .’<br />

’ I -. ,<br />

I .:<br />

. . . i+’<br />

,*!<br />

i _:<br />

procedure doeo, in fact, exist. This information wlll be valuable in<br />

facilitating in the quest for the most satisfnctcry criterion aaasureuent<br />

procedure to be used in valldatloa studies.<br />

The four rating procedures to be used a.6 criterion masure of<br />

performsnce efficiency are self ratings, peer or co-worker ratings,<br />

supervisory ratings (first and second level), and group-speciallet<br />

ratings. Theoe avaluatto~ will al1 be obtatned on fder.tlcal rating<br />

forms comprised of an appraisal of the subordimte’a “(Xterell perform<br />

ance” and a check list of his overall job-proficiency quallflcatfon_s.<br />

-_<br />

These data when collected, vi11 be analyzed in vhole as vell as by<br />

subparts to derive the most meaningful information from them.<br />

These have been just some of the more pertinent research projects<br />

conducted at EEC. Uany other research projects have been done in the<br />

p-t, and many more are now in the developnrent ataga. It is believed<br />

that only through dellgent und carefully designed research programs<br />

such as these will improvement in our evaluation syntem be realized.<br />

120<br />

,


.<br />

.<br />

Test Construction Procedures<br />

wILLI\c’ .,.. ‘: tP*?rE, ..a CIIA,.NN ,<br />

, US &my Enl;fir?d Evaluation Center<br />

The te5t construction procedure is an especially crucial one in<br />

,.obtainino, a valid instrument, especially for test6 such as o*.ir~ t’nat<br />

usuaily cannot’ be given experimentally prior to operational use.<br />

For evaluation activities to be most effective, they should consjs:<br />

of the best possible techniques, used In accordance with what we know to<br />

be the besr and most effective psychological principles.<br />

One feature that distin&uishes reputable *hotk in test development<br />

from that of the mas.; of self-styled “test constructots” or “test experts”<br />

a>d outright quacks is that the reputable worker in the field is continuously<br />

concerned with testing, verifying, and improving: the adeauoc)<br />

of his procedures, He, knows that he does not know all t!le answers, and<br />

he is cvcr on the alert to find out more and to improve his procedures.<br />

There is no easy road to scientific test construction.’ The’roat’ is loni;<br />

and tortuous and beset with many pitfalls.<br />

Xn our types of testins programs, there usually is no test available<br />

that corresponds satisfactorily tp a function h+iich seems important to<br />

test. We as test psycholoI:-.sts and subject-matter experts, and coordinators,<br />

are then truly put upon our own mettie to originate improved patterns of<br />

test performance and to develo? a crude test idea into a practical and<br />

reliable testing instrument. This constitutes the most exacting and, at<br />

the same tfme, the most interesting and rewardfng phase of tesK development<br />

work. It requires truly creative efforts.<br />

The topics to be presented by the four symposium members are some&at<br />

diverse in nature and should stimulate our thinking in the direction of<br />

improved test construction procedures. At this time, I would like to fntroducc<br />

the fou; symposium members:<br />

Mr. Isadore J. Newman, 6570th Personnel Research Laboratory,<br />

US Air Force<br />

Mr. John Crediford, US Naval Examining Center<br />

Mr. Charles E. Cassidy, US Army Enlisted Evaluation Center<br />

Mr. Fred B. Honn, US Army Enlisted Evnluation Ccntcr<br />

Since the four ?apers to be presented by these gentlemen arc somewhat<br />

diverse and to keep our thinking “warm” relative to the particular topic,<br />

we will have a short question-and-answer period following etch presentation.<br />

121 I


Educators nnd trr;incrs erc cor,tinuelly kin:: chnllcn$ad to project<br />

knov1cdge.s that en&le ir12ividunls fo not only txijas: co ti;cir cnu<br />

viroment b.lt 9llstcr i t . ‘Jnfortunately ,tb.cre ar.2 notable dtf fcrencee<br />

among some educators and trttinera in their rspncfty to perceive that a<br />

challenge exists. xnfonration is so rapidly accumulating t’nnt it la<br />

imperative that we seek new methode of dissemination, Since 60 much of<br />

man’s diverse behavior is a result of lcsrnfr?g, we must find CI system<br />

by which vcrlffcd facts and relationships may be projected to hisi in e<br />

systematized effort,<br />

Educators end treiners must seiect the cethod of bnatruttfon vh


.<br />

.<br />

*<br />

.-a . ..” .-. ,. ., “.. _... .,... “. _..._ “__e._.- *...+... . . .- ..^ _.a, .~f-*-i-Y. “. ,C^.. 1 I .h.<br />

sarfaa of taaka which their mission requires. The conxzands develop their<br />

owu measuring instrumenta to evaluate their o*m on-the-job training<br />

progra36, Our unlt has the r;Iselon of dcvelcping the meeauring instrument<br />

for USAF; b%ich, under the new conccptr will .cNJaoure the knowledge reeulting<br />

from studying the subject matter of the ca:ccr development courses.<br />

Where&a in the put we verc chnrgcd with the reeponeibility of r,;aauring<br />

cn aiman’s knowledge of his entire GP+a-ialty, we vi11 bc limited to<br />

measuring only that knovlcdgc tiich is boee.‘. on the mntcrial concerned<br />

with principlea and fundamentals found in the career development ccursc.<br />

The Air Force Training Cocnond people who monitored this ncv concept for<br />

the Air Force, have ststcd “Hodificatione of present methods will be used<br />

for testing knowledgea learned and tikills developed. Specialty Knouledge<br />

Testa (SkTe) will becorce in effect ‘end of course’ tests for airmen<br />

completing a career develo;tment course. The test will cover only the<br />

content of the self-study natcriale. The writers of course materials vi11<br />

assume a nev role ~+ich will require more corefully planned approachca to<br />

the dovelopncnt of effective coureee, Specialty Knowledge Teats will<br />

adhere to what the student h-s been presented in hie coursa. If he<br />

hoe learned the material well, he should be surceesful in passing the<br />

SAT.” With this background I CD projecting the theei that a nyatsms<br />

approach might be utilized in the hole training technology into which<br />

our test construction process might ‘be integrated as one of the crub.<br />

systems,<br />

T%e Air TraloLng Coonnend has described the development of a 6yStcmR<br />

npproach as one that ‘I.... views the mnny in~;ividuals and groups de*dcloplng<br />

a particular veapons syetez a8 individual coD+W?nt6, like cogs in 8<br />

machine working togethar to achieve a ccmnon goh~.’ This approach requires:<br />

(1) the definition in precise terms of each person’s job; (2) a task<br />

analysis; (3) a specificat:ion of performance requirements and tolcrfince<br />

limits; and (4) a statercnt of the necessary interactions and cmnications<br />

to be carried out be:rleen groups --each requirencnt established to meet<br />

the predetermined system Eoal.” (Oflesh, 1964)<br />

In this frameuork,the Air Training Command (a subsystem itself, in<br />

overall training and evaluation) require8 a systems approach to include<br />

a task analyst8 as the second step in the sequence. To be useful for the<br />

test construction process these tasks must be stated in measurable bahavioral<br />

terms. A criterion test rmst then be developed co that a<br />

starting point in the training may be dcrermincd. The training materials<br />

are then started at thia point of departure and are carried forvard to<br />

the desired skill level. Our unit then takes over as one subsystem to<br />

construct a measuring instrument which purports to measure the knowledge<br />

which the examf.noe has acqufred at a specific skill level. This subsystem<br />

ranks the students according to how much knowledge relative to other<br />

students he has acquired, the resultant being a percentile score.<br />

123<br />

,<br />

.


Th? systems concept to be mortt effective, demands a alngIC mancger 1<br />

to align the subsysths so that each nerrhoe vlth the other aa lt evolves<br />

into tha einglo eyaten. Affter KQ US&’ has ccrtlficd that the contents<br />

of a opeclfic Specialty Dcocrlptlon correlate significantly with the<br />

relevant tasks involved In the fob, Air Training Commd break@ the job<br />

down lnto tasks and level6 of proficiency vhlch rare ueed as u Job Trafning<br />

Standard, If these tvo documents, villch wa coneldor 8uboystems, are<br />

atgnlflcont, then the subsystezs charged vlth vrltlng the training materials<br />

la off to n good start in making its contrlbutlon to the total reyatm.<br />

Assuming that each eubsystes has been properly constructed, and la properly<br />

coordinated with the other subsyeteae, vork ln the SKI tooting subsystem<br />

ohould proceed amoothly since ve would then be charged only with conatructlng<br />

an instrument that memore: how ~011 a:n airman haa mastered the<br />

metcrlals found in a career development courao. With the syetcm working<br />

optfmall)i wltheach subsystem making Lts proper contribution there need<br />

be no queetlon or concern over &at the SKT ia meaouring. Under euch<br />

circwatnnccs it will be meaaurlng the objectlvce and crfterfa eetabllahcd<br />

by the subsystem charged ofth prcparlng the career development course.<br />

At this point it must be remembered, however. that the SKT ia only sampling<br />

rcpresantativc areas of a teak as datcxmlncd by the teat conatructlon<br />

rubaystain.<br />

ihccaria (1563) hna eaid in dlacuoelng the problem of maeursman~.<br />

“Too often students are meoeurcd to fraction8 of a percentage pctnt agalnet<br />

other students vlthout ever being raeesurad agelnst mlnirxnn job requirements.<br />

There are two t :ln reusona for this phenomenon. Flrat, tralntng oblectlvco<br />

ara seldom atated in definitive enough terms, and aacond, a relative<br />

rather than an abeolute measurement oyetem i8 employed.”<br />

Thlo leads u8 to a diacuseicn of the usea for which an evaluation<br />

ir made. It la important to the whole system that this use be aperiflcally<br />

announced so that it will be one of the objectlvee for each subsystem to<br />

keep in mlnd title making their contrlbutfon, If Dr. Zaccaria’s crlterlonbaeed<br />

evalua:lon isuacd,it will result in a certain group being found<br />

proficient without knowing hov proficient. If the percentile rank lo<br />

used, ltwlll tell us the relative standing of each lndlvlduel in the<br />

group. It 1s up to the eyrtem manager to decide vhlch evaluation<br />

correlates best vlth the use for bhlch the measuring instrument is dosigned.<br />

A system8 approach vi11 work oniy when each subsystem la working<br />

towurd the rame goal under the direction of a slngla manager.<br />

124<br />

--... . ._-_. ^. --- - - .----.-- - -- . _ _ __- -._-- _..,.-.<br />

.<br />

. .


Ceoly, W. D., 6 Cratn, J. L. Dual concept of on-the-job train?ng. -USAF- Instructora’ Journal, 1, (Hr. 1). July 1963, 13.<br />

Draorel, P. L. Evalwtlon procedures for gantral education objectives,<br />

@xcatfonal record, April 1950, 97-122.<br />

Rflgard, t?, R. Theorfas s learning. New York: Applaton-Century-Crofta,<br />

1956, ix.<br />

Judy, C. J. Achieveneat tenting in the Air Force. USAF Inrtructora’<br />

JOWla',, I, (Nr. l), July 1963, 17. ’ -<br />

Ltndgulat, E. P. Educational meaeur&ent, WaahFngton, d, C.: Mericm<br />

.----- -<br />

Couoell on Education, 1951.<br />

Mayor, Sylvia R. Raeemch on QuStRnrtad training at Eloctrooic6 Syotea<br />

Divirion. -111- Trends in pro-rcmmed -<br />

AosocFation, IYbb, 149.<br />

- inrtruction, National Educrtton<br />

Ofbooh, C. Air Tratnfng Comand’r System8 Dtvalopental Approach to<br />

Inatructioml Matertolo.<br />

Vitolr, 8. H. 6 Newman, L. J. A Cmpariron of Two Inotructlonrl Mthodo.<br />

Paper presented at Our Lady of tho Lake Collage, San Antonio,<br />

Tezaa, 1964.<br />

Zaccaria, #. A, Reappraisal of achievesent medeure#. USAF Inotructoro’<br />

Journal, r (Hr. 1). July 1963, 73.<br />

125<br />

,


Sucimary of<br />

Pragmatic Creativity in Excanlnntion CdnbtructLon<br />

A Paper Delivered by<br />

John Credifotd<br />

us lava1 E?mafaing cancer<br />

Hr. Ctedfford spoke on the pragmatic utilftatfon of purely divergent<br />

thinking of nonprofeseionnl item writers. Because of a change in the<br />

time schedule, his prepared material was suzxnarlzed ~6 follows: After<br />

dcacribing the f-CA aa one of the greatest potential force8 in today’s<br />

teatinS, a ploa was made for the deliberate 1ntroduct:on of opportunltiea<br />

for free-flowing thinking into our tottinS ettuotfon. Several teeto were<br />

.‘aacrlbed that wete the results of placing nonmilitary item vritera in<br />

tne position of creating their ovn teets in a purely permissive environment,<br />

Learning by doing, uohuqered by tho rcntrfctlone of<br />

claeeical precedence, they produced sevaral noteworthy teeta, including<br />

a music test on tape administered tk* :r.zmar school chilfiren, a nilitery<br />

eupervfeory test, a queatlonnalrc used in oeiactlng brig guardo, and a<br />

tast for the selection of nonbissed. s*Jp+rvteory personno ir. indust,y.<br />

In line with the top:c of the dfacuanfon, the grectcet bendfft from the<br />

pepar was in the very fluent discussion Mich followed.<br />

126


:.<br />

‘ci<br />

� �<br />

�<br />

�<br />

�<br />

i<br />

I<br />

1<br />

I<br />

. j<br />

.<br />

.<br />

.<br />

.<br />

EvaluatFoll o f zotor s:cs11s<br />

A hOmc;y Little anecdote help6 to clarify my objectives in this pc-per.<br />

My father was born in Lrelantl, ad I recall a story he told me n~ny years<br />

ago about n fellow “greenhorn” whose working hours required him to retu:n<br />

home in the darkness of late evening. his cour6e led h1.m through a p a r -<br />

ticularly dark and deserted district in uhlch anyone with a lively Irj.021<br />

imaginntion could envision the direst of misfortune6 taking plscc. Every<br />

night as he passed through thie area, he constantly repeated aloud,<br />

“Praise the Lord and the Civil ain’t tl bnd man either.” In like vein I<br />

would like to pay due rcspcce to -the written test for its past contributions<br />

to the field of achievement testing and if.8 prospects of even greater<br />

utility in the future, but at the 6ame time, to recognize that the t.eRtfng<br />

of mOtor skills does offer Q great potential to be explored in our<br />

constant effort6 t0 create incree?ingly more effective measurement<br />

instruments.<br />

The rvaluatfon of motor skills will be treated in R q*~itc brosd<br />

COntCXt. The covernge includes all activities pertelnfrq co or involving<br />

muscular movement, not to exclude those requiring previous, concurrent,<br />

and subsequent cogniti*ve procesecrr.<br />

In effect, performance testing in<br />

ita mo6t conprchcncive appllcntion ~111 be considered. All test eituatfono<br />

in b-hich the examine& is rcqulred to do something Other than take a<br />

paper-and-pencil test: will be included. The use of driving tears as a<br />

prcrequisftr forobtaining a driving license in many titetea ir B good in-<br />

dication of the w!de public acceptance of performance test6. Although one<br />

might question the validity of these test6 BS typically administered,<br />

there con be no doubt that n driving test ie a pnrticularly good exqle<br />

oE a test Uf motor 6kille.<br />

The purpose cf subject-netter or achievement tcstfng normslly fr to<br />

provide an evaluation of the level of fob nascery attained by the examtnee<br />

in a given Job, and frequently tc rank o group cf individUala in regard<br />

to their relattve succe68. Depending upon the nature of the activity to<br />

be meesurcd, and frequently certain extraneoufi restrictiona, one test or<br />

R battery of two or more typca of tents may be ueed. In the selection of<br />

measuring instruments we place cnphaeis upon the p:esence of three b8sf.c<br />

qunlitiee: (1) vnlidfty, cr measuring what w want to measure, (2) reliability,or<br />

consistency of tTICiiBIJrf?ment, and (3) objectivley, or the erc-<br />

elusion of personal feeling not based upon accurate observation. We have<br />

at OUr disposal a variety of measuring inetrumenta*<br />

Llritten tent - The multiple-choice ~ypa of written test hoe become I<br />

almost: univerxy accepted ns the basic meonure. It poesecses the i<br />

127


, *<br />

-. aa*<br />

Y<br />

. -..<br />

. . ’<br />

advantages of very complete coverege of the appropriate subSect matter<br />

on the verbnl level in a comparatively short time, simplicity and euee<br />

of administration, relatively IOU COBt, SimplC scoring proccdusee, und<br />

finally, it fs suttable for prqsenting problems that m?esure many cypvpco<br />

of abilities.<br />

Performance rating - This consista of an evaluation of the individual’s<br />

performance on the job by his supervisor; in the Army it is called the<br />

CEZR or Comandcr’s Evaluation hcport. Althcugh ratings of this nature<br />

are susceptible to subJective elements, the design of the reting form<br />

attempts to direct the rater’s attention to obJectfve observation of<br />

behavior,<br />

Pcrformnnce test - fifs mq be in the nature of a work umple such<br />

a8 a typing test, a tapping test to measure an aptitude such as finger<br />

dexterity, or a sitoational test in which actual vork problems nre aimsla&d.<br />

Certain ae?ection and promotion boards can be properly considered<br />

performance tests as can the selection interview in an employment office.<br />

This is true when the interviever or board is attempting to evaluate<br />

b?havlor required in the performance of Job dutiee.<br />

Evaluation of experience and trnu - In 8ome goverzticnt Jurisdictions<br />

a scnre is aiven to each applicant for selection or promotion<br />

bneed upon quality and quantity o.f jbb pertinent experience, iraining,<br />

and special recognition such as awards received. Thie wore become8 part<br />

of the examines’s final rating.<br />

It is obvioue that all Job; require some basic motor activity. Certain<br />

movements of the feet, hands, eyes, and other parte of the worker’s<br />

anntsmy are required to reach the L3.srk aituntlon, to position himself for<br />

the performance of his duties, and to control and manipulate the physical<br />

tools of his trade. The absolute requirement of some degree of motor skill<br />

is present from the most sedentary of occupations to those thrnt require<br />

almost constant motor activity. For our present co?zern, that of testing<br />

military personnel in their occupational apecfaltiee, the broad spectrum<br />

of Jobs is divided into three classes. This purely arbitrary taxonomy<br />

based upon relative importance of motor activities in discriminating<br />

between levels of Job mastery provides a starting point in the determination<br />

of the need for A test of motor skills. Class 1. A large group of service<br />

personnel are required to perform duties that require only the basic<br />

motor skills, the posoeseion of vhfch can be assumed from their<br />

cacceptunce into the service. Although M)tor coordination contributes<br />

something to the efficiency with which they perform their dutles, Its<br />

importance fs ovtrshedowed by the fmportance of cognitive functions which<br />

are central to the executfon of their normal work requirements. Administrative<br />

and general clerical occupations are exsnnplee of this group. It<br />

appears thst a pcrforoance test would contribute little, if anything, to<br />

the eveluetion of these personnel. Class 2, A large number of eervice<br />

. .<br />

.<br />

I.28<br />

,<br />

c<br />

! . *,!<br />

,)<br />

.


.<br />

personnel are assigned to jobe the duties of k%lch require rotor nkil.ls<br />

of 8 mre specific nnture thsn those in the previous category. ThCQC<br />

skilile till have usually been acquired in one of tha service achoola, or<br />

in some casea, will hove been poosesscd by the individual upon entry into<br />

the scrvicc. The motor aspect of these skills might contribute oubstantially<br />

to the quality of the performence of their job duties, but<br />

cognitive abilities are of central importance. The ability to smoothly<br />

nanipulate tools, position physical objects, and perform manual movements<br />

in the adjustment or operation of equipment is required, but the evaluation<br />

and diagnosis of the situation requiring these physical skills le a better<br />

determiner of competence. In this erea, the quart ion is vhether tbe<br />

central skills can be musurad adequately oy a written test, or should<br />

some form of performance test such as the performance check list be<br />

made part of :ho total evsluation. This decision must be made by the<br />

test psychologist vorking cooperatively with subject-matter expcrtn in<br />

the epproprlate field. Before arriving at 8 final decision, pertinent<br />

validation data nnd frequently experimental c’cita must be considered.<br />

Various types of repairmen, mechanics, and equipment operaeors are exsmples<br />

of personnel whose jobs are of this nature. -Class -3.<br />

A COFparntlvely<br />

small number of jobs exist in !&ich motor skills appear to be<br />

crucial discriminators between levels of job mastery. These positions<br />

are typified by duties that involve const.ant, rcpctitivc activities that<br />

lend themselves readily to qt;an::itlve and/or quolitatlve measurement.<br />

Horc important than the fact that these job activities are readily<br />

measurable, is the knowledge that studies of pc. formance on the job<br />

compared with scores on the typical wrftten tests have often indicated<br />

that there is little correspondence between the knowledge of vhat to do<br />

and the ability to do the wrk quickly snd accurately. For joba in this<br />

category a completely adequste measurement of job mastery must include a<br />

performance test. Typist and sterographcr are examples of jobs that are<br />

Included in this class.<br />

The decision as to bbether or not the evaluation of a particular job<br />

should include a performace tcet rent6 entirely upon the nature of the<br />

elements of the job and thelr susceptibl~lty to meaaureneet by a written<br />

test. This declslon can be made by the test psychologist only when<br />

complete job analysis lnfornatioa fa available. ‘I’he analysie of the job<br />

provide8 a liar of duties &ich are required for adequate job performance<br />

and must serve as a starting point in the determination of the types of<br />

tests appropriate to constitute a complete evaluation, The actlvftiee<br />

must be defined, analyaed, and elements necessary for job success must be<br />

Isolated. This process involves the dlvialon of the job into its basic<br />

elemental components, a process which some Cestaltisee will find objectionable;<br />

however, if a particular element of a job fs not adequately<br />

,<br />

129<br />

.<br />

.


-<br />

‘. .<br />

*. .<br />

;-<br />

. d<br />

--<br />

. . .<br />

.<br />

. I<br />

_.-. .’ - . . .<br />

.<br />

.-- __ _ . ._<br />

measured by a wrltton test, we are challenged to find A way to evaluate<br />

this factor. We can concern curaelvca with emergenta, whan WB have<br />

conquered the conetituenta. Dasic criteria for job 8ucce~a euch a5<br />

quantity of output, quality of output, accuracy, spoilage, and eafcty<br />

factor8 muot be determined. Decfslom mat be roade a8 to whether the<br />

performance will be neaeured in terns of pro&Act or process. If it ie<br />

to be product, should WC establish standards rrncnoblc to objective<br />

ueatiureaent such us size and weight, or will subjective mcasureti such<br />

aa moothnees, color quality, and aymetry serve as better doterninera<br />

of competence? If an emphasis i5 placed upon proceae, should we be<br />

primarily concerned with use of tools, proper work methods, or work<br />

aequencef The selection of universnlly superior work procedurea f~ not<br />

@Any. Should a mechnnic vho con consietently analyra correctly a mAlfunction<br />

8imply by listening to the en@& be penalized because ho<br />

doesn’t use the generally accepted tool5 and work nethods? Thie ie only<br />

one cxaarple of the type of problerna to be solved in the construction of<br />

a p,ood performance teat: but it ie a good indicntor of the obstacle5<br />

that a teet conetructor must hurdle.<br />

Aptitude - ~+is is A test of a rotor skill that attcmyta to predict<br />

auccea5 in a particular activity or potential for benefiting from<br />

training. Aptitude tests are not Job achievement tesia and are mentioned<br />

here only because they are one type of performance test and in 8ome<br />

situation5 can be of considerable value. Ihe distinction between attitude<br />

end achievement teeta Is not alway cryetal clear. The U&Z! of the ;Cet<br />

rather than ite n.turc io the important factor, The UBQ of a aubjcctmatter<br />

teat to predict success in a position of a hi.gher level ie closely<br />

related to the u5e of the typical aptitude test.<br />

Achievement Teets<br />

Work aample - l-hi5 test provide5 the examinee with a typical. performance<br />

oituation appropriate to the Job for which he la being evaluated<br />

including a teak or group of tacks characteristic of that required for<br />

actual Job performance. A work srmrple ie not actually a piece of a Job.<br />

Some part5 of any Job would yield little atatiatically u5eful variation<br />

in performance; other part8 might not be adaptable to a teotlng eituatfon.<br />

A good work aampla muat differentiate betvesn good and poor workcre, rind<br />

provide 5core8 reflecting degree8 of proficiency. Thie is poseiblo only<br />

if a fatr 85mpl@ of crucial determiner8 of job 5ucceaa 15 included, requiring<br />

the cxaminee to demonbtratc hia acquired ukilla uofng the tools,<br />

maCerfaL5, tnd method5 characteristic of hfa Job. Teats of this type<br />

have been developed and utilized in the EEC for typieto, stenographer8,<br />

bandsmen, court reporter6, and radio code receivers.<br />

Situational performance tests - Thoro test8 do not attempt to pctttsure<br />

a eimple activity, but one !&rich is rather complex and lea5 well defined<br />

and isolated than the work sample. A group oral teat in which such<br />

130


. . _-. .._. -...a_-e,-,. .-Y.-e--.. .-V.-d_ -..w-.... -..-- ._-. - .I. . . --.<br />

I<br />

. .<br />

131<br />

/<br />

.___,__ .__- -. . .._ -... . -.-.+ - .._, .I .<br />

personality factors as dominance, leadership, judgment, and emotional<br />

6tnbflity are evaluated is an example of this type of te8t. Probira<br />

eolving ability ha6 nlso been mesaured in thfa way by prerznting the<br />

examinee with a unique simulated s:tuation of the nature he might encounter<br />

on the job and rating hie hendling of the situation. Tne<br />

process test, which ia primarily concerned with proper vork procedure8<br />

and sequencea, and the perfonaanct? check lilt are also typec of the<br />

situational pcrformnnce te6t. Hr. Claude Bridge6 of EEC 5.3 presenting a<br />

paper at 1330 houro thie afternoon in the Vest Auditorium on the aubjcct<br />

of perfbmance check lists. Because cf the complcxitiee of administration<br />

and acoritg, high cost, and the availability of a variety of written<br />

tests covering the appropriate eubject matter, the eituational performance<br />

test has not been widely utilized in achievement tearing,<br />

Performance tests for evaluating murk proficfency should be uded<br />

only when a group multiple choice test cannot provide en adequate measure.<br />

In nany instances the mastery of the job con be inferred from the fact<br />

thnt the individual possesees eufficient kncwledge to perform hlc job<br />

duties. Certain jobs, which by their nature, are centrally concerned wit-h<br />

a re;)ctitive, manual activity require a performance teet to supplement<br />

the written test in providing a complete evaluation. A typist, for<br />

exzrple, might have. a good knowledge of the various parts of the typewrfter<br />

and their function, but experience haa shown that this knowledge is not<br />

highly correlated with typing ability a8 measured by a typing performance<br />

test or performance on the job. Jobs of this nature represent only a<br />

relatively small percentage of jobs in the military services. The most<br />

fruitful area for further research appears to be that of the situational<br />

performance test, or more specifically, the performance check list. As<br />

a starting point, more comprehenefve fob analysis is required, job<br />

element8 predicting GucceGG must be defined, the procedural verauo the end<br />

product problem must be rccolved, and better method8 of scoring muet be<br />

developed. The old bugaboos of increnPed expense, greater expenditure<br />

of time, and difficulty in providing adequate and equitable tett sites<br />

are still with us. The competitive group oral and situational problem<br />

solving test6 offer encouragement in evalusting certain hard to meatiute<br />

pcrsonnltty traits; however, their principal value will probably be in<br />

ccrt8in specific unusual teat situations rather than in the area of job<br />

proficiency testing.<br />

__ . - . . . .- --- --- /-<br />

,<br />

.


.<br />

Perfonnsrnra Teat Construction<br />

US Army Enlisted Evaluation Center<br />

Having sntcbliahad the need for a performance teat, we must firat<br />

gfve due conrideration to the objectives of all perfo-ncn teat-e. &r<br />

prime objective is to measure the levels of proffcisncy Fn those critIca<br />

skilla which are not measured adequately by estsblfshad written teaca.<br />

Our Becondary objective ia to deattnifne If the oxamincae achieves (demonotrateo)<br />

skilla which are important, or critical. to mfnfmal parformanca<br />

on the tarkr of tha job. Thnt Fa to nay, we must identify motor (ccilnurl)<br />

rkfIlo that are to be sampled by our parformrinca ta@t in chess categories:<br />

firrt, thora skilla which are critical Fn dlatlngufrhiap different lavalo<br />

oL job marcery; aacond, those vhtch or-6 eeaantial for accaptwblc performante<br />

of the motor trek8 of the job, and finally, those skills which cannot<br />

tanaon@bly be measured by written teats.<br />

Our first step fo to identify which Job taut8 rdquiro thaern rkills<br />

previously mantionad. Prom such A very cmprobcnefva study of the job<br />

tarkb we ~111 be able to pinpoint the actual test trsk. Think of the teet<br />

task a8 our roed F&p which tall8 ua how to get to where we P~Q going.<br />

To a~~ure that our analyrfa is complete end accurate, we ur? the<br />

tack analysir method. Thie cormonly entails the crccompllrhmtnt of a form<br />

utillxlng fivn columpa. Tha hsgdtngo for those colunne are: (1) ACTIVfZP<br />

STEPS (what the worker actually doca); (2) PilOCEPURPI (how the taskr muat<br />

be done); ;3) CAFE AKD USE OF ‘MATERIEL (the toolo and tqufpaent with which<br />

the tmekr are done); (4) SAFETY AND SPECLAL PR!XAUTIU?-JS: and (5) COXUlSIoNs<br />

(reeultr of havfng done these tarks, or work samples). Performance tc8ts 5re<br />

ured for several purpoosn. Providing a criterion measure Fr one goal for<br />

performance testing in tha anied 8CrviCcO. The UIIQ of the task analysis<br />

method ir the beat assurance we have for achieving this goal.<br />

Once thir a~lysir bar been completad, v4 are in a position that<br />

� nablea US to ~QO the whole picture. Hence, we can more readily identify<br />

critical pointr such aa, not measurable, not a ratable point, or to<br />

label A point a8 vary critical. Actually, the taok anslysLa (S~Q prefer<br />

“vork uamplr”) bacomca :he OOUTCQ of items, ratable points, for 0p.r<br />

chxperimsntal inatrumant. -Further, thir analyrir qrortly aide o u r aelection<br />

of the correct type of test inotrument. Will ft be a final product,<br />

a procQII, or even a ccmbinatfon of product and.p ocerr test?<br />

132<br />

I<br />

- ,. ._......____ - .__- ----<br />

.


The octond 6reA of opercatfon’ for performance tetrtero le the conetruction<br />

of the experimental instrument. Au we @tart the proceoo of<br />

&electing rateble points bared upon our AcYI~sIE, there cara et leeet<br />

three criterle which should guide our selecticn. The firot criterion is<br />

reprerantativenecr--the activitice, being measured are rcalietic (joblike),<br />

typicA of the taakr and okillr performed by A quAlifi.ad apectatiat<br />

on the actuA1 job. Performance tenters Are frequently tempted to aubatitute<br />

unraaliattc talka, for a variety of reasona, and often cacrifire<br />

all cr almat all of the criterion, repreaentativenasa. (3318 10 not A<br />

fccetioua refarence to honeat recognition teats or aimlated conditions<br />

tcrte.) Our second criterion le reliability--conaiotent rerulte yielded<br />

for a rerpecteble range of acoree --enough so that a standard can be<br />

developed for world-vide wage. Third, va tast conatructorr muut be<br />

mindful of the criterion, prcctlcality-- feaoible to u&e in tome of the<br />

time, equipment, parosnnel. und ex?enrc to adminirter and to @core.<br />

Raving relected the “itaml” for our eicperimentel instrument in<br />

accordance with the three criteria ve muet next develop the rating<br />

scale needed to evaluate the tasks to be maarurcd. Banically, tie hsve<br />

two typeo of acelee frcxs which to chooat+: the forced cholco;go/no go,<br />

did/did not or the degrees of skill, vhich can take any of oaveral forma,<br />

Auch � 0, numerical (1, 2, 3, 4, 5): deccrlpttvo word scale--for exempia,<br />

poor, good, &VOrAg%, etc. A poerible third acelo, phyrlcml characterirtics<br />

o f the final prodr.:t, ie aomkthea uoAd. Remaber, i t ir eaey<br />

to go overboard with numericel ocal~)s, AI K. L. Ecan (1953) pointe out.<br />

Only thooe rater8 vho are exceptfonally well qualified ohould attempt<br />

uoing A rcale of more than five pointr. The nature of the items of our<br />

inotrment will largely dictate which ocale ir more appropriate. A<br />

word of caution at thfr point Fe in ordar. We tmut ba careful to piAn<br />

porrfble weighting procedurer vlth an eye to chcckfng them egninrt our<br />

crltarle. Ao A rule of thumb, complicated, involved weighting rhould<br />

be avoided. Hany promising teats have been invalidated by cumberroma<br />

weighting.<br />

Our next mc\Jor area of activity f0 the preporstion of inrtructionA<br />

for the axamlnerr And the examtneer. Basic contsnte of the exaninee<br />

fnrtructionr should inc?.ude A deocriptlan of the tank to be done, A lirt<br />

of the pointe of the tetik that will be scored, all the tool8 and/or<br />

equipment to be used, � ���� lkaft (ff applicable) and any other necclaoary<br />

lnatructionr.<br />

Huch more detailed inmtruc:ionr are necessary for our examinerr.<br />

Not only murt we cover the darcription of the taako, ths materiel list<br />

And time limlta bur alro tip8 regarding careful obaervatfon of tha<br />

examfnea’r ptrfomnce--eopeciAlly prwetior?t for 8 btep vhich fr, or<br />

could be difficult to obeerve. Further, theae inotructionr must be<br />

explicit regarding the recording of reruLts with the rating scale. we<br />

P<br />

_ _._ -_.- ._-._ - _-. ---<br />

,


.<br />

.<br />

kxmws hat’ to do if the axamin~a makes<br />

, OR undo en alternate method, OR etarte<br />

ult in fnjury or dmz.aga to equlpmsnt.<br />

hi8 quota from Adkinr, at al (1947):<br />

capcafed by axplicit inotructfonr to the<br />

ratinge and what to look for.”<br />

f. The next activity in ths expsrbneata’l<br />

ctore uoe the etandord teat-reteot tachnique<br />

E fnrtrument. If applicable, rateeting<br />

a helpful. A critical point to rcmmbor<br />

arga -age of examinean--from Lhooe who<br />

s to the most capable onea. (Ideally,<br />

es would be divided into two groupr-y.)<br />

We muat be aura to check and rachack<br />

ru of our instrument. Rcaearch haa shown<br />

ter8 can pay big dividends in .the area of<br />

ility.<br />

our tryout include hslp in doter-mining<br />

the tert; holp in datewining the mfnirxn<br />

sskfng for the Amy. wa would be helped in<br />

:-rrder’s Evaluation and the ET (written<br />

tryout rorsults most ruroly will lead us<br />

the inrtrumont and, poaoibly, even eme<br />

tryout takear us a @lentic step toward<br />

10 into gripe with rcalf*tic analyefe and<br />

Thir necerrltates coneiderable coordlnatfon<br />

:ara end field pcreonnel, not to mention the<br />

I. The purpore of this thorough coordination<br />

reffne our efxperFmsnta1 Fnotrument until<br />

, reliable tast inatrumant. We will have<br />

!dmLnlrtrotions of the official instrument<br />

knd elfmfnats problane not pravfourly antlciidard<br />

forme, 0coritlJ keyr, manual8, etc.<br />

.rtfcal analyrLs of teat renulta In order<br />

Fnstnment for use on a world-vida baafo.<br />

.<br />

H. L., Brfdgeo, C. P., &<br />

of - --I echievment me tc@te. Waeh-<br />

; Offlca, 1947.<br />

i peraonnsl tests, Wew York:<br />

uant. iiarhingtoa, D. C.:


O&#&RAL%YWP011U0111l-SRRS.GE~SEVIF.VESCHULTER,CEIIAIRB;1AW<br />

us HAVAL EXAbllltZ:Q CEIPTER<br />

One Interpretation of the Mjor<br />

Coals of Specialty Knowledge TcatirLq in the<br />

United Stateo Afr Force*<br />

ST8P¶iEN w. FcmS<br />

6570th Peroonnrl Renearch Laboratory, US Air Porca<br />

Kaybe itDo an occupational disease afflicting home of ua who rbork on<br />

the aeocmbly line grindit% out the tests. But it tiometimee happens that<br />

we get so wrapped up in the daily routine of the job and BO preoccupied<br />

with the short-run goals that we’re apt to lone sight of the major goals of<br />

the task. We bscoam no involved in the workaday methods that we find it hard<br />

to urecramblo the enda from the means. So, from time to ticnz, it is worthwhile<br />

to climb aloft and take a fresh look at our referenca poit?ts and renew<br />

our perepective. It ie occalrionally necessary to forget the mode of travel<br />

and concentratr on the destinntion.<br />

IN SUPPORZ OT THI! A?CRMN CXASSlCFIC%TXON SYSTEK<br />

The expreesed goal of the Air Borce specialty knowledge testing program<br />

la to evaluate the technical knowledge possesued by nirran a8 required for<br />

qualification under the Air Porte enlisted personnel classification eyatem.<br />

Towards thie p,oal, the Specialty Knowledge Test (XT) is provided aa an Air<br />

Porce-widd etandard of measure by which to determine job knowledge, apart<br />

from job performance as 8~1~. The SKT is applied not only laterally by<br />

career specialty but also vertically by skill level--namely the apprentice<br />

or eeniskilled, the journeyman or ekilled, and the advanced levels of qualification.<br />

Ae a criterion for skill upgradinS, the SKT ie intended to supplement-but<br />

not supplant--other criteria, such a8 demonstrated proficiency on the job,<br />

job experfence and history, supervisor’s recorrenendation, and coaxzander’a<br />

approval.<br />

upgrading.<br />

Thus the SKT is by no mean8 intended to be the eole criterion for<br />

Toward8 managerial control. In effect, the SKT aervea a8 a managerial<br />

control device whereby Headquarter8 USAP ie enabled (1) to ensure that the<br />

airman manpower resources meet the oetabliehed minimum requirement6 in terms<br />

*It is emphasized that this is one individual’e interpretation. Thio u83er<br />

doer not neccssartly reflect the official policy of the Air Force. No; hoes<br />

it necessarily reflect the position, whether official or unofficial, of any<br />

major air commend. Appreciation Is eepacfally due Lt Co1 Albert S. Rnauf,<br />

USAP, for the stimulating dialogue that evoked many of the observations noted<br />

herein.<br />

.<br />

,<br />

136<br />

L<br />

_ . -- ._<br />

i<br />

_... * ,<br />

f


.<br />

.<br />

.<br />

of technical knowledge, and (2) to accwlish a rza~ure of sttndsrdtrat?on cf<br />

airman knowledge on 6n Air Force-wide boeis.<br />

Oualitv control<br />

_L - - • It is axioostic thut knowledge ie power. Eiobherc is<br />

the truth of thin axiom better founded than in a tmdsrn military organlrstion<br />

I in which the succc~o of the nfseion depends on the qualitative superiority,<br />

r&that than quantitative strength, of ire F;E.npower, Knowledge, then, may<br />

be regarded a- a form of resource. Like t’lZteria1 reBturce8, technical knowledge<br />

, io subject to deteriorntlon, Leepletlon, and obsoleacenrc. To be n-aintaiced<br />

i.n a conetant state of readiness, knowledge rnuot be continuouely reneved,<br />

restored, and cultivated. Unlike material reBource6, however, knowledge is<br />

not readily mnable to rneneuremcnt for inventory purpoaee. Nevcrtheleas,<br />

througi. the SKT program, the Air Force seeko to maintain a slose check on<br />

the job knowledge reaourcea that reside in the enlisteJ nanpowcr population.<br />

Gtandarditation: one AP language.<br />

-I Through etandarditatfon of knowledge,<br />

the Asorce seeks to neutr$tllzt the hazards of specialization and division<br />

of labor. One hczard is the propensity of each organization todrvelop its<br />

own concepts, its own doctrine, ita own private language of epeclalized<br />

teminologv and nomenclature. One could fancy the crnergence of an Air Force<br />

tower of babe1 as the thec,retical ootcorae of this tendency, carriei. to it6<br />

ridiculous extreme. However, it should not be unrealistic to credit the<br />

SKT with making a notable contrfbution to the CLPUEQ of standardization of<br />

knowledge. Acting aa a vital stiolulant to the currency of a coamn technical<br />

language, the SKT helps keep the Air Force family of comorande on the cm<br />

wavelength for purpoeee of concnunication.<br />

Standardization: the “c~leat” air.-c!an. Another hazard is the tendency<br />

for each mjor air con-x to foster that job infornntion that la directly<br />

relevant to it8 own mission, to the virtual exclusion of broader areaa of<br />

knowledge having wider applicability to the general Air Force miseion. This<br />

is understandable inasmuch as each command ia under relantles8 preaaure to<br />

meet the demands of ite irrnediate mission. On top of their many burdene,<br />

the commands bear the donkey’s share of tralnlng burden. If an individual<br />

does a creditable job towards the fulfillment of the conmvlnd mission, the<br />

commend will naturally tend to want to overlook any knowledge gaps that might<br />

limit the individual’s potential value to the Air Force at large, As long<br />

as he ia a good SAC man, who careo if this man could ever be of any earthly<br />

or airborne use to MATS, or vice versa? Well, Headquarters USAF cares, And<br />

. so should MATS care from the very standpoint of ite own long-range intereete.<br />

And 80 indeed should SAC. For the atrman’s breadth af knowledge stamps him<br />

at once with both his professional and hir, Air Force identity. It is the<br />

. mark of his versatility and employability within his specialty. l’hiti quality<br />

in the airman largely relieves the Air Force of the need for retraining him<br />

extensively with each change of assignment--whatever the command, whatever<br />

the fob, whatever the specific nature of the equiprent involved.<br />

*<br />

137<br />

.


.<br />

Systematic upgrading. At the same time, -the Air Force achieves other<br />

gains through the use at the SKT PS a managerial control device. By channeling<br />

the upgrading proce56, tile SKT regulates the flow of technically knovledgenblc<br />

personnel up a host of career laddere. Rclpero, apprentices, journeymen,<br />

tcchnici6ns, 5upervisor8, superintcr&nts--all arc kept advancing at on orderly<br />

pace, each group maintaining the preecribed distance of feparetion from the<br />

othere, A kind of braklng mechanisu, the SKT kcepo ampetition from disintegratin!<br />

into a chaotic ecramble and providcn a hedge againot the runaway Inflation of<br />

ratlnge. A sort of traffic control system, the SKT set8 rz&xm speed limit8<br />

in the form of qualifyfng percentile scorea.<br />

Wotwithetandfng limitations. To be But@, the teat doe8 not have the effect<br />

of sunnnarily and irrevocably eliminating in drove8 failures from any furthercompetition,<br />

On the contrary, through reteating and board action, all but a<br />

negligible proportion of the failure8 eventually get by. Fhat ie important,<br />

howover, 18 that the test doe8 serve to keep the rate of progression within<br />

manageable limit8 for purpose8 of effectrve personnel salectlon. And to be<br />

sure, one can envisage a c-are flexible application of speed llmite, depending<br />

upon what the traffic will bear. A rather permis81ve puss/fail ratio might<br />

be justified for a critically undermanned specialty, or a very restrictive<br />

ratio for an overcrowded one. However, :.xh manipulation to accomdate variable<br />

supply-demand relationships azxng rpecialtics would hevc to be carefully conoidered<br />

in thr. light of poaeible conflict with the goals of 5tandardfzat:on and quality<br />

control,<br />

All told, the resulatory effect of the SKT function cannot be denied. It8<br />

value a8 a absnagerial control device ie appreciable, elrpecially when a88eEsed<br />

against the alternative of no control.<br />

k’or the good cf al.l. The gains achieved by the SKT program are by no<br />

means confined to the hig!lest level8 of management. The beneficieries of the<br />

program am found at all levels right down to the individual airman. Couxnandere<br />

and 8upeTvi5or5, faced with mekfng selections from amxtg relatively homgeneous<br />

group8 of personnel, find A trusty catalyet in the SKT, helping to ease the onus<br />

of decision. Heanwhile, the individual airman is provided with an objective system<br />

of career progreoslon, giving him the opportunity to compete in a eervfcewfde<br />

arena. Under etandard conditions, the testing situation Afford8 him the chance to<br />

demonstrate anew his capacity, in term8 of specialized know-hgFI, for advancement<br />

in level of responeibility and authority.<br />

IN SUPPORT OF TRAINING .<br />

Thus far,the discussion has dwellod mainly on those goals of the SKT program<br />

that relate to the use of the test a5 a managerial device in support of the airman<br />

claeslfication system. There are certain other goals of the SKT that produce an<br />

impact on the training function.<br />

138<br />

.


#<br />

___.<br />

-<br />

: ..\ \<br />

c<br />

. .<br />

I<br />

. i<br />

I<br />

1 i<br />

.<br />

.<br />

.<br />

.<br />

.<br />

Screening byppasa cpecfalloc~. The use of the SKT for the identification<br />

of bypass specialists is ohe of these goals. A recruit, &IO has sme ~acialty<br />

backgourd derived from civilian ocoupational experfencc or schooling or fros<br />

prior military service, azy t&e the SKI to ascertain his qunlificntion in the<br />

spec‘lal ty. Cm passing, he is svarded the smtskflfed rating,. Every year the<br />

bypass-specialist program yields significant savircge in term of cfrc-umented<br />

trainfug costs.<br />

Hotivatinq study. There is another goal of the SKT progrran that brings Lt<br />

into an affinitive relationship with the training program. This goal is i-licit<br />

in the publication of study reference lists for the guidance of alrum in preparation<br />

for specialty testing. The effect is to motivate study on the part of the<br />

airman-to kindle his urge to acquire the technical knowledge that is considered<br />

crucial to his successful pcrfo,rmance on the job and to his long-range career<br />

development.<br />

Lrpact on Iraining. As a consequence, preparation for the test become big<br />

businese. &my people get into the act at all levclo. Training program go<br />

into high gear. Tne impact upon training is felt throughout the Air Force, but<br />

nowhere is it felt more acutely than at the Air Training timand, In turn, the<br />

latter undertakes to produce training standard,0 and job informtion for b.ir Porte<br />

publication. r;ot only are these made available to tha traLnee for his use in<br />

preparation for the test, but they ate also adopted by the XT progrfun for input<br />

to the test-construction process.<br />

A r.c*~ epproach to CUT. A noteworthy outcome of the interaction between trainin;>.<br />

and evaluation has been the recent adoption by the Air Force of the dualchannel<br />

concept of on-the-job training. This approach to CUT. provides for the<br />

synchronous development of the airmen’s cereer, on the one hand, and his job proficiency<br />

on the other. The end product of one of these training chaonels is the<br />

Career Development Couree (CCC), which is a self-study course geared to the<br />

aiman’s specialty for his use in preparing for the next higher skill level.<br />

Since it is a self-contained peckage of career specialty information on fundamentals<br />

and basic principles, the CDC is a welcome source reference for use in<br />

SKT construction. Thus has the SKT fulfilled itself, in part, through ite<br />

salutary *act on training.<br />

The isaue of fndependant evaluation, So salutary, in fact, has bean the<br />

two-way interaction between training and evaluation that, for practical purposes,<br />

a sort of symbiotic relationship has emerged betveen the two. The intimacy of<br />

this relationship has been the subject of considerable interpretation and, possibly,<br />

overinterpretation.<br />

-I_ The ca6e against. One strongly voiced interpretation holda that evaluation<br />

should be an integral part of training. Accordfng to this view, the SKT program<br />

would logically be assigned to the Air Training Cormmnd. It ie maintained that<br />

the closest possible coordination 1s needed to effect greater efficiency and<br />

economy of test production. Such a wedding would eupposedly enhance the mutually<br />

supporting relatfonehip between the two functions. Thus a higher degree of mutual<br />

responsiveness would become possible In a more intimate association.<br />

139<br />

;<br />

i<br />

,i


2<br />

. .<br />

. -<br />

“.,.>__ ., . _ _ _ _ . . . . . _ _ -<br />

4.h.. ._ _ Tha caBe for. Another view, the preveillng ona, is that eveluation should<br />

ram&in independent of treining. Evaluation, it is contended, is properly a<br />

coordinaea function vfe-a-vie training and should not be relegated to a s~ordinata<br />

role, To do so%uidx BornewhAt like Qlscing the bar examiners under the<br />

low-school faculty, the teacher-certification agency under the normal. school,<br />

or the Auditor under the bookksepar. tiir would ignore tha nosd for checka and<br />

balances between training and evaluation A% indcpend-antiy coexisting functions,<br />

As an evaluative instrument, then, the SKT is supposed not - -to<br />

bs a form of<br />

training de-rfca. To regard the SK? 8~1 euch would supposedly be a groea nioconception<br />

of it6 true function, even though the SKT dcee have the extrinsic<br />

effect of operAting As a ntudy LrcpeIler.<br />

I a’<br />

/<br />

* \.<br />

.’<br />

,<br />

--<br />

c<br />

According to its Qroponents, evAluatfon is better able to respond flexibly<br />

to rapid technological developments with the Introduction of new subject matter,<br />

And to cope with the sudden obsolescence of old subject matter, eince it is A<br />

far simpler mtter to revise a test than to revise a training courne. To submerge<br />

evaluation in tralnfng, it is pointed Out, would impoea upon the former<br />

the inherent drsubacke of the latter. It in considered cruciA1, therefore,<br />

that evaluation remain directly recrpooet.re to the clsesificntion stclndards<br />

rather than become elavishly dependent upon the errining standards. As for<br />

the ecorrcmies anticipated with the pro?oaed consolidation, it le believed ehicl<br />

would have to be calaulatsd in term of overell cffactivenecs rather than<br />

purely Dionetary units. The sacrifice of IimSted pecuniary aavi~~ge is be-ieved<br />

to be a relatively cheap price to pay a~ part of the coot of independent<br />

nvAtuAtion.<br />

Theee two opposing positions retative to independent evaluation form the<br />

horns of a dilemm that periodically rear into view. ‘What’8 the eneuer?<br />

Whatever the Ansver, it mst first be recognized on both aider that tha<br />

qussti.on iteelf is recilly neither one of tralnirg nor of evaluation in an<br />

exclusive eense. Rather, it is a question of rcanpower utilization in A<br />

corcpreheneive renee and probably should be Approached Aa such. Whatever the<br />

answer, it should be identifiable neither as training policy nor a8 evaluation<br />

policy but as a amnpower management Qolicy. In short, it should be an Air<br />

Porte anmer to an Air Force question.<br />

Surnnary. And there It ie-one vereion of the mjor goale of specinlty<br />

knowledge testing in the Air Force. Ye have reviewed the goals of the SKT<br />

as an instrument of management in support of the airman pereonnel classification<br />

ayetern through quality control and standardization of knowledge and through<br />

systernstfc career progreeeion. We have alao reviewed the goals of the SXT<br />

in relation to training, not only to energize study but also to screen bypass<br />

epacialiste. Finally, we have scrutinized the icpace of ehe SKT on the training<br />

program And raised the question& independent evaluation, In 00 doing, we hnve<br />

euggested that, whatever the answer, it should be transcendingly Air Force in<br />

it8 spirit.<br />

. .


.<br />

,<br />

.<br />

.<br />

-<br />

.<br />

.<br />

Uses of NOS Evaluation Teot Results<br />

J. E. HMREITM<br />

US Army Enlieted Evaluation Center<br />

Dr. Bruner, a mathematician and geophyoicist, more renowned for hie<br />

contributions to hi8 fields than by the number of his publications, once<br />

asked his students to differentiate between mathematicians and the colculur.<br />

After a few minutes of their profound, evasive silence, he pointed out<br />

mathematicians use calculus to solve problems which neither can solve<br />

alone or which mathematicians can laboriouely solve with lees efficient<br />

methods. Test6 of occupational capability, like the calcu?u.s, are tools<br />

which can aid in the solution of per?onnel management proceasea only when<br />

they are capably used. TeGt users, like mathematicians, can more<br />

effectively solve their perconnel evaluation problems &en they make full<br />

and appropriate use of the tests available to them.<br />

The fir-et Army HOS Evaluatior Tests were edmlniotered in January 1959<br />

as a basis for the award of proficiency pay. They were called HOS Proficiency<br />

Tests and, for the most part, s-led the abil?ty of enlisted<br />

personnel in grades E-4 and above to recall the fundmcntols of their<br />

primary HOS training. The raw test 6corea vere converted to Arm; standard<br />

scores, weightnd, and added to weighted Commander’s Evaluation <strong>Report</strong><br />

rating scales co provide a composite “proficiency score.” The te6t acores<br />

. .., and proficiency acores were reported to the enlisted personnel concerned<br />

and their unit personnel officers on a form entitled “Proficiency Data<br />

Card.” Summaries of teat results were furnished to Headquarters,<br />

Department of the Army, and major cousnands. The minimum score for the<br />

award of proficiency pay for each Hilitery Occupational Specialty (!%3S)<br />

was determined by the training requirements and attrition rate for the<br />

MOS and the number of proficiency payments that could be made. Lists<br />

of minimum proficiency score8 were distributed throughout the Army for<br />

corrnnanders to use as a basis for individual proficiency pay awards. Ae<br />

you can readily see, the first MS Evaluation Tests (MOS Proficiency Teets)<br />

were used in two vaye: They were used by cormmurders and unit personnel<br />

officers to determine which of their enlisted personnel met or exceeded<br />

the minimum rcquiremcnts for the award of proficiency pay. The tests were<br />

also used by enlisted perrronnel and their officers to estimate how they<br />

ranked with all others tested in their MOS.<br />

Current trends in the Army MOS Evaluation Program are tovard increaeed<br />

emphasis on: sampling enlisted personnel’s LJilitieo to solve job<br />

problem, multipurpose scores, and improved reporting of results. Becauee<br />

of the wide variety of specialties in the Army and the large number of<br />

personnel tested, realization of these trends is necesearily more complete<br />

for some NOS than for others, though some progress has been made in all<br />

141<br />

.*


.<br />

-<br />

. . -.. ..- . .-. .._-,_.-&s -..-._..-..... a-.-r-..< .I.- -w--w ,, -- . -_ ..-. . ..- _ -...-<br />

nreat3, Aa new ldcns arb developed, or m6thoda of adapting old fdeeo are<br />

visualized, lmnediate goal5 .wi.ll ,ba extended and trcndb: m’dified accordtngly,’<br />

Since the account of things to came. mat evolve from the pteeenc f the scope<br />

of thie prcaentatfon io limited to currant hippcning~ with a faw indices<br />

of next steps md anticipated hazardo.<br />

The emphaafr in teat covarage haa been shifted from trafnfng materi<br />

content to the problemc encountered by pereonnal senlgncd to the HOS<br />

between their graduation from formal training md their prowtforr to in<br />

advanced ski?.1 level or MOS. Par HOS in which cepcciaily needed (Eandsmen,<br />

Typists, S:cnographere, Court Repotters, and Radio Code Operators) performonca<br />

tests of motor and sensory skills have been developad for<br />

standard adminietration world-wid6 and are used a~ supplements to papcrand-pencil<br />

tests to product more job-related composite ecorea. Job-semple<br />

problems, adapted for multiple-choice anawerfng, huv6 been developed and<br />

uRed in aeverel testr. (These problems are bneed upon repreaentativa<br />

aeaignments for the specialty skill level with coannonly encountered<br />

aituctlonal data presented in nnrrattve form, or recorder tapce, or<br />

drawicge, fully or partfully completed forma, or ocher vieuel meana.<br />

Representative eolutfono of outstanding and infer-lor specialiats are provided<br />

for choices.) Several test outlines have been radevelopcd along<br />

f!:nctional linea to produce eubncores which reflect the comparntfve<br />

abilitiee of examineee to perform the various duty pcnitionp within a<br />

sp6Cialty. Atl can readily be seen, the net result ie a ronvsreion of the<br />

primarily job knowledga tests developed in th.1 early stages of the Army<br />

Evaluation Program to more predcrinantly job-pI’oblem eolvfng abilitiee<br />

tests. When the converafon ie complete, all KOS Rvaluatfon Teet. ecores<br />

can be used with confidence to determine how well examfneee ten perform<br />

the current mafor dutiee of a military Bpecialty rat.her than how much the<br />

examlncoe knou about the fundamentals of their KOS.<br />

Tvo additional index scores ere developed and reported: the NOS<br />

qualiflcatfon acore and the promotioa qucliftcatfon acore. The KOS<br />

qualification score le a *minimum passing score used to determine vhich<br />

sxtminees should be retrained or reassigned to P mor6 appropriate<br />

rpecielty skill levsl. The score can be used to determine which examfneee<br />

with a critfcal primary MO5 will not be awarded the sllpplemental<br />

pay for their assigned opeclalty. The method for determining the HiXi<br />

qualification score is bacied upon the premise that experts can arrive<br />

at the absolute number of question8 in e teat which, if answered correctly,<br />

distinguieh between the minimally qualified and unqualified pereonnel<br />

aaefgned the H03. The Promotion Queliffcat:on Score nay be used by<br />

commandera, if they desire, es a requirement for the advancement of enlieted<br />

personnel within their commend to a higher pay grade or ckill level.<br />

It is the score attained or exceeded by one-third of the examinees arai.gned<br />

in the same pay grade in the HOS skill level teeted. Where such comparfaone<br />

142<br />

_.-


- .<br />

.<br />

‘..<br />

!<br />

_<br />

. I<br />

1.).<br />

--<br />

.<br />

.,2<br />

‘.<br />

L’<br />

.<br />

.<br />

:<br />

. I<br />

..-<br />

. .<br />

\<br />

_-<br />

.’<br />

‘s . . ,,<br />

‘.<br />

i<br />

c *<br />

:<br />

4<br />

\<br />

/<br />

:’<br />

. 1<br />

.<br />

.<br />

were made, this score comnly 1.~38 In the range of avcrcgr! wares made<br />

on the s&ne test by soldiers in the next higher pay grade.<br />

The Proficiency Data Card has been replaced by the WS Evaluation<br />

Data <strong>Report</strong> which, In addition to the MOS Eveluotfon Score, reflect8 on<br />

a five-point stole the proportion of the itcza in each QajGr test creil<br />

the examinre answered correctly. NOW, all exazzfnees and their 5upcrvioors<br />

have available a medium for determining how each excninee’a compooLce<br />

Bcore (test + commander’s rating) compare5 with the Bcores of all other<br />

Active Army exminees with that primary HOS 61~111 level and how each<br />

exazzinee succeeded or failed in answering the questiocq related to the<br />

major teot areas. Examinees can and should uBe their reported test scores<br />

88 a guide when preparing for retesting during the next scheduled evaluation<br />

period for the NOS. If they concentrate on improving their skills In the<br />

areaa In which they answered the snaller proportions of item8 correctly,<br />

they will broaden their overall MS ca-pahflltLes, and usefulness to the<br />

Army, more rapidly and efficiently than they can by divcrsftying their<br />

study cfforte. Supervisors and unit coxnanders can also identify the<br />

test areas in which their subordinates answered the smaller proportions<br />

of Items correctly by reviewing and summarizing the XOS Evnluction lIeports<br />

of their oubordlnates. The results of their reviews can and chouid i.!<br />

used to plan their training programs and training emphasis.<br />

Each major coz#nnnder is provided smary report5 of the teat acorcs<br />

attained by the enlisted personnel tested within his coumarnd. These reflect<br />

the distribution af the MS Evaluation Scor.e of the personnel tested<br />

by the comnand distributed according to NOS skill level, pny grade, and<br />

principal coaznand subdivisiona. From these reports the major conmondera’<br />

staff can determine: (1) the numbers, locatione, and pny grade5 of personnel<br />

tested who failed to attain the minimum score for the avnrd of a<br />

verified primary HOS; (2) the numbers, location, and pay grades of personnel<br />

teeted whose scorea are in the upper third of the acoree of all personnel<br />

assigned the primary WL!~ skill level tested; (3) the numbers, locations,<br />

and pay grades of the personnel who may be awarded specialty pay; (4) the<br />

numbers, location, and pay grades of the pereonnel who may be awarded<br />

superior performance pay. Theee reports supplement the strength reports<br />

of the command by providing information concerning the capabflitfee of<br />

personnel assigned enlleted specialties within the conmwd( The staff<br />

is no Longer limited to reports that there are X men with specialty Y<br />

In the command, The command twmary reports add: A of the X men are In<br />

the upper third of all of the Y speciallets In the Army; B of the X men<br />

failed to cylnlify for a verified primary HOS; C scored above any determined<br />

more point; the average score of personnel assigned to a given specialty<br />

ekfll level was , etc. Such data can be used for estimating training<br />

needs, locating specialists for critical aaeignments, and related administrative<br />

processee.<br />

143<br />

.<br />

i


During the post year, the fiOS Evaluation Tisting Program has been<br />

extended to the evaluation of the ability of Active P.my personnel to<br />

perform :he duties of their sccondery HOS and the ability of Reserve and<br />

National Guard personnel to perform the dutiec of the N0.S for thefr current<br />

duty positions. Uhile the fame tests and reporte are used for thece new<br />

purposes, evaluations are beaed upon teat acores without courrander’e<br />

evaluation ratings and are used only to identify those exmtneee who<br />

failed to attain the Active Army minimum qualifying test BCOI’C for nosignment<br />

and training purposes.<br />

fumy l?OS Evaluation Tests are also used in determining the pay grede<br />

and HOS of commissioned and warrant officers vho intend to enlist cr reenlist<br />

upon termination of his current active commissioned or varrant<br />

officer service, Eligible off tcers are permitted to take aa manv as three<br />

HOS teat6 appropriate to pay f rade E-5 or their prior enlisted temporary<br />

grade, if higher. At least c!-e of the tests must relate to an HDS which<br />

hue vacancies at their reqtleh!ed pry grade. None of the tests may involve<br />

an overstrength MOS. Finol tictcnninntion of pry grade and MS ia mnde by<br />

the Department of Army Grade Determfnetion Board based upon the teat<br />

acore6, cotprlander’s recommen;ations, and other pertinent, available<br />

data.<br />

In each of the current gpiicatl.ons of ;.rmy HOS TeLting Programs, it<br />

is the connander who must take the final action. The cocszander can give<br />

an enlisted man tiho fails to make a passing score and acquire a verified<br />

primary HOS a second chrnce to improve his skills before the next<br />

scheduled test session ior the HOS or reassign him to a more appropriate<br />

MOS and okill level, but must reclassify the man who fails twice, The<br />

connander can withhold Ruperior performance or opecialty pay from personnel<br />

who attain an eligible score if he determines the individual is not currently<br />

performing in a satisfactory manner. And major commandera may restrict<br />

promotion8 to those whose scores fall within the upper third of the acores<br />

for their specialty skill level.<br />

Several have euggested that MOS Evaluation Teat reeults be ueed to<br />

evaluate the effectiveness of service school training. At first glance,<br />

the proposal appears to have merit; but closer inspection reveala crucial<br />

f allaciee. In the first place, eervice school training is ordinarily<br />

designed to provide enlisted personnel with the basic vocabulary, theory,<br />

methods , and procedurea that enable them to begin their on-the-job training<br />

at their first unit of assignment. However, HOS Evaluation Teats are intended<br />

to cover the period from the completion of on-the-job training to<br />

advancement to a higher akill level or MOS. In the second place, the<br />

students vho fail the school NOS course are rarely, if ever, aoaigned the<br />

NOS where they might succeed or fail depending upon the degree to b%ich<br />

school training met on-the-job needs. In the third place, many of the<br />

144


enlisted personnel evaluated, particulnrly senior tcc!lnicfans and noncommissioned<br />

officers, have not attended a service school course for many<br />

years, if at all. Consequently, much of the material acquired during<br />

service school :raining by such personnel b.>uld not be covered by teuts<br />

restricted to curxent doctrine and matcriai. It should ail&o be obvious<br />

that the highly-motivated, rapid learners k&o do uell in school courses<br />

usually perform better on the job and NOS tests than slow learners and<br />

those not motivated to make the Army their career. Any correlation<br />

between schoo! grades and MS test acores in influenced by the quality<br />

of school training and by lear,ling abilities--motivnticn factore. The<br />

degree to which positive correlations are increased or decrascd by the<br />

quality of school training and its relation-hip to MOS requirements<br />

cannot be determined from the school grades--test score correlations<br />

directly, Additional studies would be required to determi:;e the extent -<br />

and direction of the influence of training upon job success, While<br />

there are probably other reasons why MOS Evaluation Test scores should<br />

not be used to evaluate the quality of service school courses, the four<br />

reasons offered suffice to negate the proposal.<br />

Others have suggested HOS Evaluation Tests be used to determine<br />

whether persons called to active duty in a mobilization may be assigned<br />

directly to a unit or whether they require preliminary training bc fore<br />

assignment to a specialty. The Reserve --National Guard testing program<br />

obviates the need for retesting of members of Reserve and National Guard<br />

units who will be assigned to Active Army units or issued current Active<br />

Army equipment. But HOS Evaluation Tests cannot be used to determine the<br />

abilities of reservists, guardsmen, and draftees to operate, maintain,<br />

and employ the limited-standard, demothballed ,laterial ubed in situations<br />

requiring major mobilization, because MOS Evaluation Tests cover only<br />

current, Active Amy doctrine and material. It would be necessary to<br />

develop additional tests of the abilitles of exeminees to perform the<br />

specific duties of the additional specialties to be used in the mobilized<br />

force and to establish minimum standards of performance for those<br />

specialties to determine which examinees require further training and<br />

which could be assigned directly to a unit. Tests of basic theories and<br />

principles would only rate the abilities of examinees to learn--not to<br />

do. It is possible that time limitations in major mobilizations would<br />

preclude the development of adequate placement tests and that the benefits<br />

derived from such a program would be less than those resulting from the<br />

refresher training of those who pass the tests.<br />

.<br />

Some have also suggested MS Evaluation Tests can be used to predict<br />

how well an examinee will succeed on the job. That is, how will his peers<br />

and supervisors regard him? Certainly, one must agree that such a goal<br />

is desirable, However, the question is: “Is it practical at this time?”<br />

145


I--<br />

,<br />

c<br />

Now, an MOS Evaluation Test. rates examinees’ ebilitiec to perform the<br />

full scope of the duties of an HOS skill level, weighted to fit the<br />

objective3 and concepts of relative importance of the standards setting<br />

level of the Army pro&ream manager3 staff. In other words, the testo are<br />

biased in favor of the broadly-skilled examinee rather than one who is<br />

exceptionally skilled in a pnrt of his HOS skill level reqafremcnts.<br />

The test score indicates the exazinee’s relative standing among all<br />

examineee ansigned the MOS skill level. If peers and supervisors can<br />

adjust their ratings of exr=,inces to comyenaate for their personal<br />

bioaes and influences, if they thoroughly understood and accepted the<br />

objectives and important concepts of the standards setting level of the<br />

program manager’s staff, and if they could simultaneously and comprehensively<br />

evaluate the abilities of all personnel assigned the primory<br />

MOS skill level, they would probably rate the exnminee as the test rated<br />

him, if the examinee reacted to them as he reacted to the ceet during the<br />

evaluation period. Ohviot~sly, peers, auperviuore, and exainees interreact<br />

differently, differ in their opinions a8 to the relative importance<br />

of tasks, have varying degrees of understanding of the specific objective3<br />

of top level standard setter3, and have rarely, if ever, had the opportunity<br />

to evaluate the abilities of 311 persons assigned to any MOS skill level,<br />

let alone nimultaneously. The test3 do provide.a basis for predicting how<br />

well an examinee could do on the ; Jb when permitted to do the whole job<br />

rather than a oubspecialty and when motivated in the acne degree and<br />

direction aa he was at the time he took the test. A greet deal more information<br />

concerning his interpcreonal relationships and attitudes along<br />

with those of his peer3 and supcrvisor3, their rating attitudes 6J d<br />

abilities, and the dynamics of the group would bc required before one<br />

could predict Low an examinee would be regarded oy his peers with<br />

reasonable accuracy.<br />

A few item writing and technical publications writing groups have<br />

cooperatively used test question response data as a cue to whether technical<br />

material has been dietribuced to and understood by examinees. With<br />

few exceptions, the test question analyses developed by the Enlisted<br />

Evoluation Center reflect how all of the personnel in each MOS skill level<br />

responded to each test question. Items relating to equipment or doctrinal<br />

changes are carefully reviewed when only a small proportion of the CXaminecs<br />

answer the questions to determine whether the queetione or the<br />

cxaminees are deficient. Checks are also made to determine when guides<br />

and training materials were distributed to exeminees and whether these<br />

material6 require clarification or amplification.<br />

Many will envision other u3es of PiOS Evaluation Tests. In planning<br />

the uses, one must bear in mind and insure:<br />

a. The te6t content and objectives are compatible with each<br />

planned use.<br />

trolled,<br />

b.<br />

All pertinent variables are identified, evaluated, and con-<br />

____ __- .--.--- --<br />

146<br />

-.-_- _._. _ ..-.._. -<br />

- - -_ - -.


,<br />

.<br />

c. Test-ocorss which only reprtaant uslotive rtandlngr of<br />

%x~~~fncer ore not logically summed, multiplied, aubracted, or divided<br />

by ordinary arf thnotic p~xmmt1s.<br />

d. The cmponmto and wcighttngo of cmqcsbta ecoreo era<br />

cczqatible with the plantlad u8e.<br />

e. A teat deeignusd for one purpose cannot noceesarily ba<br />

used fox whet appear8 tc be n related pcrpoee.<br />

147<br />

;;<br />

^f


f<br />

.<br />

Job P.na1ysi.r for Tart Development F’urpozars<br />

Prank H. Price, Chsi-n<br />

US Amy Enlisted Evaluation Center<br />

Today the grmt need in mfXitary teiting -- perticultrly job proiiciancy<br />

evaluation -- is a sound understmding of the baeic job. Gnly<br />

if the job to be evaluated lo known in all of itn detailed charactoristics,<br />

can teata be developed to adequately measure euccess in the job.<br />

fn other wrdn, we naad information about the job in vhich ue hope to<br />

ta8t proficiency.<br />

Job onalysia la a proteus of obtaining information about jobo. While<br />

fob an~1yri.s can serve vorioua uoeful peroonnel purposes, the most Important<br />

one for our consideration ia that of retting pereonnol rpecffications<br />

required in a particular job. For tent construction purposes, a job<br />

dercription is the end product of job analysis. It ir vital that the job<br />

dercription be complete fn every detail. Success or proficiency in a job<br />

cannot be properly wetueted unless tho nature of the job is fully knaan.<br />

There era a ambar of different agproachnm to enalyxing jobo and<br />

writing derr:iptiona of them. The papcro which Dr. Horah and Nr. McBride<br />

will present this morning mre illurtratPve of different approaches to<br />

obtaining job Infonaation. We believe you till find the papers intereotlng<br />

end inform.stivo; and we hope thay vili stimulate job anslyrir effort In<br />

your mm organIs~t1one.<br />

148<br />

-- --. . - c<br />

,<br />

.<br />

.-<br />

.


I .<br />

.<br />

. _ ,-,:. ._ .._ ., -.. --.. “., __ _. * * _<br />

.,. . .,<br />

New Perrpactives in Job Analysis<br />

JOSEPH E. WRSC<br />

6570th Personnel Research Laborctory, US Air Force<br />

. .<br />

As the result of an intensive research program during the past ftve or<br />

six years, the United States Air Force has dcveioped and applied a novel<br />

procedure for collecting, organizing, analyzing, and reporting comprehenafve<br />

job lnformatfon. The procedure cmbines features of the check list method<br />

with those of the open-ended questFonnaire and the observation interview<br />

into a single integrated procedure. I am certain that research findings and<br />

products obtained thus far have implication s for proficiency test development<br />

beyond rhoae that have been utilized. I piopose, therefore, to discuss the<br />

meth:? in some detail and to present some typical end products in the anticfpatfon<br />

that the potentialities of the method will be provocative of fdeae<br />

and will elicit from members of this symposium suggestions for future research<br />

and computer progrerusfng.<br />

Advantages of the Air Force Method<br />

The Air Por-.e method of job analysis has a number of advantages over<br />

traditional methods. The procedure LB simpie, economtcal, and flexible.<br />

It makes feasible the survey of large samples. It fs based on joint responsibility<br />

of job incumbents, test .:ontrol officers and unit coaxnanders. The<br />

job information fs obtained in etandardfzed , qusntffied or readfly quantifiable<br />

form. The information is current and has been found to be highly<br />

reliable.<br />

Job Anal*.rsis Operatfons<br />

The Air Force job analysis procedure involves a sequence of several discreet<br />

steps.<br />

a. Location and procurement of source materials.<br />

b. Construction of first draft of job inventory.<br />

c. Interview review of first draft by technical advisers.<br />

d. Revision of first draft of job inventory.<br />

e. Pield reviev of revised draft by senior incumbents.<br />

f. Construction of operational job inventory.<br />

g. Selection and location of survey sample.<br />

,<br />

149<br />

.


h.<br />

i.<br />

1.<br />

k.<br />

1.<br />

m.<br />

n.<br />

.<br />

Reproduction and mailing to eelected ‘X0’s<br />

Administration of the job inventory<br />

Responding to the job inventory<br />

Receiving, scanning, coding, and collating<br />

Key punching and verifying job inventory data<br />

Electronic computer analysis of frurvey data<br />

Distribution of survey remits<br />

Now let us look briefly at each of these steps.<br />

Source Materials<br />

The source material8 used in the construction of job inventories coneist<br />

of the specialty descriptions in Air Force Hanuals 36-l and 39-1, Job<br />

Training Standards, On-the-Job Training Package Programs, Training Course<br />

Outlines, <strong>Technical</strong> Orders, and any other pertinent publications. A reference<br />

library facility is being built up which provLdes current source<br />

materials pertaining to all airman career fields.<br />

Construction of First Draft<br />

An Air Force job inventory covers tasks performed by all skill levels<br />

of one airman career ladder from spprentice, through journeyman and supervi8or,<br />

to superintendent. Three persons work together In constructing the<br />

first draft of the inventory. A personnel technician or job annlyst select8<br />

duty and task statements from published source nateriale. Upon his judgment<br />

the quality l>f the inventory largely depends. A clerk-typist prepares successive<br />

drofte of the inventory and may derive preliminary task statements<br />

from selected sections of publications. A supervisor editor checks forn;nt,<br />

wording and organization of tasks statements into duty categories and coordinates<br />

the development of related inventories. Construction time for a<br />

job inventory varies with the complexity of the career ladder. For the less<br />

technical ladders, three to four weeks is adequate for writing the first draft.<br />

For the nore technically complex career ladders the period may be twice as<br />

lows -- six to eight weeks -- OP even lor.ger.<br />

Interview Review<br />

Fran three to eix technical advisers who are in the appropriate career<br />

ladder and are usually experienced senior NCO’a, are interviewed individually<br />

or as a group to obtain their constructive critfcism of the first draft of<br />

the inventory. These consultants are frequently the same subject matter<br />

specialiete who arc assigned on TDY to the Personnel Research Laboratory to<br />

build Specialty .Lowledgc Tests.<br />

150<br />

I<br />

_<br />

c<br />

.<br />

.


I<br />

,<br />

. .<br />

.<br />

.<br />

(,. ..> I. “._ .__^ - ,, ..*. ., ./<br />

Bevisicn of First Draft<br />

- -<br />

. .<br />

On the basis of the suggestions and recorenendatl.ons of the technical<br />

adticers the first drraft of the inventory is revised. Teaks which urre<br />

not lisC:d are added, dutfee oad tetzkn not performed nre deleted, and f3properly<br />

worded statements are corrected.<br />

Yield Revlev<br />

- - -<br />

.<br />

UcualSy frc-m 15 to 20 copies of the reviEed draft of the job invntoq<br />

are reproduced In booklet form. Theee booklets are then malled to Test<br />

Control Officers (TCOe) at rrelectcd baecs in different gf?ogPnphiCal arena<br />

la major air commands. The Test Control Officer8 distribute the inventorlea<br />

to senior job incumbents for review. The senior NW’s, like the technical<br />

advioere, are instructed to add duties aad tasks which are not listed, Co<br />

delete tasko not performed in the career ladder, to revise improperly worded<br />

statements, to meke reccoreeadatioas for improving the invcnioq and to return<br />

their suggesttoaa and comments to the Personnel Reeearch Laboratory.<br />

Construction of @peratloaal Job Lnventory<br />

- -<br />

Tasks statements added by senior job incumbents in the field cre extracted<br />

verbatim from the invector; booklets, clnsstffad by type, and grouped<br />

by duty category. After careful consideration and close inspection for overlapping<br />

atatemente, decision 5.6 made regarding the ecccptance or rejectf.oa<br />

of each ndded statement or susested modification. Accepted tcsk statements<br />

are coll.ated with statements Fn the inveatorj under their reepective duty<br />

headings. This second revision coastitutee the first operational form of<br />

the job iaventory.<br />

Sesnple Selection and Location<br />

_I_-<br />

Sample size depends upoa the aumber of incumbents available in the<br />

career lndder being surveyed. Since 2,000 ie the limit of the computer<br />

program capacity, thfe sets the maximum size of the sample. An attempt fe<br />

made to obtain anproximetely 500 incumbents fn each of the four skill levels<br />

in the career ladder. In order to insure having a stctistically adequate<br />

sRnple of each skill level, surveys us~.~elly have not been conducted in any<br />

career ladder where fever thaa 500 airmen are aseigned.<br />

Reproduction and Mailing<br />

TCO addresses end numbers aad locatfone of incumbents in the appropriate<br />

specialties are detenined from manning information suppliee by headquarters<br />

of the several ccmmande. The aumber of inventory booklet0 to be published<br />

may vary from about 600 to about 2,500 to allow for booklets not completed<br />

for one reason or another. Sufficient copies of the inventory for the portion<br />

.<br />

k<br />

;<br />

Ii<br />

:,<br />

)-’<br />

;<br />

of the sample under hi.8 jurivdicttoo are mailed to each participating TM.<br />

Included in the package are administrative directions and other fnotrucrfone<br />

for hendling and returning the booklets.<br />

151’<br />

/<br />

I<br />

I<br />

I


.<br />

.<br />

Administration of the Job Inventory<br />

,<br />

�<br />

Teat Control Officera conduct the group ndministrntioq of the job<br />

inventory in base testing rooms. They scan completed booklets for adherence<br />

to directions and return them to Khe Personnel Research Laboratory.<br />

A tyRica1 job inventory of some 300 task statemcnta require8 about<br />

two houra adminfatration time.<br />

Reeponding to the Job Inventory<br />

Job incumbents in the selected sample complete the inventory by first<br />

supplying certain identification and biographical information. They then<br />

check al! the tasks in the inventory which they perform and write in any<br />

tasks they do which are not listed. Each incumbent’s statements written in<br />

by incumbents during the survey are transcribed, classified by type, and<br />

grouped by duty category. The job inventory in then revised by adding the<br />

acceptable write-in statements. This final revision of the inventory ia<br />

prepared so that a current instrument will be ready whenever a reaurvey is<br />

required.<br />

Key Yunching and Verifying<br />

Upon completion of a survey, incumbents’ responses entered in the<br />

inventory booklets are key punched into electronic data processing cards<br />

and verified. For each incumbent in the sample there is required a “background<br />

information” card, a “position title” card ond several task response<br />

carda, the Ku-rber depending upon the number of tasks in the inventory. One<br />

such taek reaponsa card is required for each 69 tasks In the inventory.<br />

Computer Analysis<br />

And nw we come to the phase of the Air Force job analyefcl procedure<br />

which justifies the “new perepectivea” of my title. It is In the processing<br />

of occupational data by means of the high speed electronic computer that tho<br />

most recent major advance has been made. Computer progr8ma have been written<br />

for the publication of a job description of the work performed by any specified<br />

group of individuals. These groups may be identified in terms of current<br />

skill level, grade, command, time on the job, geographical location, kind of<br />

base, typa of previouo training, or any other variable deeircd.<br />

Routinely, the statistical analysis of the occupatfonal data includes<br />

for each ekill level, apprentice, journeyman, supervisor, and superintendent<br />

cmputatLon of the percent performing each task, Aloo computed are the<br />

average percent time #pent by members of each group who perform the task,<br />

�<br />

152<br />

_ . . __- _-__ ..-. -- .._...._ -. ; _-..<br />

I<br />

.


___ _- .__. _.._-._-. - ._.........I._.___._ _ _<br />

.<br />

.<br />

_<br />

. .<br />

. .<br />

I<br />

. ,. . .<br />

the average percent time spent by all mcrmbers of, the group, both performers<br />

and nonperformers of the task. The cumulative sum of the average percent<br />

tfne spent by all mmbera of the group is also shown so that for any group,<br />

tasks that consume 50 percent, 75 percent, or any other percentage of total<br />

tima can readily be identified. Tasks arc printed out in descending order<br />

of time spent on them.<br />

. Identiffcatlon of Job Types<br />

.<br />

.<br />

,<br />

Perhaps the most important statistfcal breakthrough, however, is the<br />

application, by Bottenberg and ChcLstal,of a hierarchical grouping computer<br />

progrsm,developed by Ward, to occupational data. This progrsm, which represents<br />

a major advancement in the state-of-the-art, groups together incumbents<br />

who perform essentially the same work activities regardless of<br />

skill level, grade, experfence, or assignment. In any career ladder there<br />

are many jobs which for all practical purposes are identfcal. The individuals<br />

who do these identical jobs are oaid to belong to the sane job type.<br />

In the grouping program,the ccmputer locates frm among perhsps 2,000 incumbents<br />

who have completed a job inventory for a particular career ladder,<br />

the two individuals vho have the most similar jobs. The computer does this<br />

by comparing every possible pair of fncumbents,ln the sample. A single job<br />

description for this pair is developed with accounts for their work tfmc<br />

with the least error. The ccxsputer then tests all possibilities of combining<br />

a job description of a third irdividual with the first accepted pair,<br />

or forms a new pair. This process is continued until finally the computer<br />

forss a group consisting of o,l members of the sample and reports the error.<br />

The iterative process may be terminated at any stage in the grouping<br />

program. The stopping point la a matter of jgldgfng when the error term<br />

resulting from merging somewhat dissinflar groups becomes unacceptably large,<br />

In a study involving 836 cases in the Personnel Career Ladder in vhich 35<br />

job types were identifled,the grouping process #as stopped at the 118 group<br />

stage. At this stage there were 27 groups containing five or more members.<br />

One of these groups which was composed of ttu, groups identified earlier in<br />

the program was listed as two separate job types and seven other job types<br />

were generated at later stages in the grouping process.<br />

In some job types work is concentrated upon a few tasks while the work<br />

of other job types is quite diverse. In general, it is found that the number<br />

of tasks performed is directly rclsted to skill level--the higher the level<br />

the more tasks are done. Certain supervisory job types can readily be differentiated<br />

fras technical job types, both in terms of specific tasks performed<br />

and in terms of skill lc*Jels of members forming the groups, Other<br />

job types cut across skill levels.<br />

Purposes of Job Analysis<br />

The Af.r Force method of job analysis has been designed, not for a<br />

speciffc purpose, but rather as R general procedure the results of which<br />

153<br />

.<br />

-, .I ! ’<br />

,]: ,/


.<br />

‘,‘,<br />

.<br />

:<br />

. . ;<br />

.<br />

.,---<br />

-.<br />

can be adapted to many uses. The Personnel Research Laboratory ROV has, or<br />

can develop, programs for producing repcrts to service the needs of many<br />

agencies. In connection with training, these program can be used to validate<br />

training standards, design training courses, deteminc which tasks should be<br />

taught in school and !&ich should be learned on the job, indicate which tnska<br />

should be taught early and which should be postponed, and so on. Proyrmas<br />

can be developed to validate qualitative personnel rcqiirementa information,<br />

to aid in the establishment of specialty qvalificntion requirements, and to<br />

identify the need for new specialLies and ahredouts. Job analysis results<br />

may be used to guide the developmen t of selection and classification testa,<br />

to improve assigcxnent procedures, to determine standards of job performance,<br />

to provide basis for job evaluation, and to contribute to manpower and organirational<br />

analysis.<br />

Job Analysis for Test Develoment<br />

In addition to the purposes outlined, one of the major functions of<br />

the Air Porte method of job analysis is that of providing data for teat<br />

development. Data derived frw computer programs may be used to maximize<br />

the content validity of Specialty Knowledge Teats, and to eatablish better<br />

measures of on-the-job proficiency. Results of many surveys now available<br />

show the percent of members of each skill level performing each task and<br />

tasks performed by the various skfll levels arranged in descending order of<br />

time spent on them. AE the grouping program becmes operational, tasks<br />

performed by various job types in each career ladder will be idectified.<br />

Hew Perspectives<br />

At the present time,vigorous research efforts are being directed toward<br />

improvement in flexibility and capability of the current computer grouping<br />

programs, Other research is devoted to the identffication of significant<br />

task rating factors and to the development of methods for obtaining other<br />

ancilliary job information from incumbents or their supervisors. Many<br />

possible factors are being considered. In studies now under way, the following<br />

task or job rating factors are being investigated:<br />

a. Frequency of task performance.<br />

b. Importance of task ccmpared with other tasks done*<br />

c. <strong>Technical</strong> assistance required.<br />

d. Difficulty of learning to do taek.<br />

e. On-the-job training required to perform task.<br />

.<br />

_.-..<br />

_<br />

.--‘, --- -<br />

_-_ _ _- - -.-.<br />

~


‘.<br />

. . ‘.<br />

.<br />

.<br />

.<br />

.i,<br />

.<br />

-,<br />

‘..<br />

‘,.<br />

‘. .<br />

- -<br />

-.<br />

i I. ,,<br />

i ,*<br />

.\ 1.<br />

.<br />

-_'<br />

1(<br />

.<br />

.<br />

j , _<br />

.<br />

_. * 1 I., ..-. _.<br />

f. Difficulty of learning to do task by OJT.<br />

g. Training emphasis task should have.<br />

h. Time apent in ~pccinl training for fob.<br />

i. Extent to t.hich job give& satisfac:ion.<br />

In Borne surveys the task rating factor used in addition to time spent<br />

is specifically tailored to fit s particular career ladder. Similarly,<br />

other information sought has specific reference to certain special coI>rscs<br />

or kinds of equipment used. In a recent survey of the Administrative Cnreer<br />

Ladder, for example, incumbents were asked to indicate admlnlstrativc ccurBe8<br />

they had attended. They were also required to gfve the number of hours per<br />

week usually spent in typing, and whether a menuai, electric, or both kinds<br />

of typewrlirr were used. Since the Training C-and was interested fn indications<br />

of the words-per-minute rates to which students should be‘ trained,<br />

incumbents were asked to check their typing speed on a six-point scale. The<br />

survey data have not let been analyzed but with an fntidental sample of 105<br />

inventories the following results were shown:<br />

AveraRe typing speed<br />

Under 15 W!?X 3<br />

15 - 24 wF?4 5<br />

25 - 34 WE4 9<br />

35 - 44 WFM 30<br />

.45 ” 54 WFM 36<br />

55 WPM or Over 17<br />

Total 100<br />

If the final survey, when validated, corroborates this am;le, it appeara<br />

that the results have obvious implications for training.<br />

Great advances have been made from the traditional observation, check<br />

list, questionnaire, and interview methods of job analyrris. The feastbility<br />

of the Air Force method has been azmply demonstrated in an operational setting.<br />

Some of the potentialities of the method have been examined but for<br />

the most part, the field of occupational analysis is comparatively unexploted<br />

in the light of modern scientific techniques. However, new peropcctlves have<br />

been revealed and at leant Borne of the problems have been identified and<br />

defined.<br />

155


.<br />

CL%TiS D. l!CYRIDZ<br />

US Army Artillery and Hissile School<br />

Tha US Army i* one of tha large&t producers of achievmant type tQRtR<br />

within tha United States today. During fiscal year 1964 over 1,000 different<br />

nchlsvemmt-type teats W~J produced by the US Amy. Them schievamrrctype<br />

tests mu referred to as KEi Evnluotion Teat8 within ths US Amy.<br />

The US Army fa organized under a decanttslfrcd concept for tha production<br />

of the reqc?rsd touting r=rtcrial becauee it r;ould rtquFre a very lerge<br />

and expensive teat construction rtcff end edmfnintrotivo crtatlirhment if<br />

all the teat materiel was produced at a ceotrclieed point.<br />

The US Amy Enlisted Zvrluation Center, Fort Benjamin liorriaon, Indiana,<br />

6CtO QO ths coordinating agency foi the ?lOS avaluatton testing program<br />

between the varicue rchoolr with the US Amy tie fumioh the r&w test<br />

mtarial and the Office of Perumnel Operfltiona, Department of the Army.<br />

When 8 dacantr~lfred orgtifaation of thie type in used in the productfon<br />

of test mter’ol, aavctsl problem ere created for US A:=y terting<br />

peraonnsl. One of them pioblm mea&!, MO3 Evclulttfon Teat outlimB, i8<br />

the topi.c for our discueolon.<br />

A teat plcn or by US Amy tsminology an MOS Evaluatfon Tcet outline<br />

for a opacific KOS Evaluation lert ia originally conrtructed by test .spacialietr<br />

at the US Army Enliatcd Evaluation Center- Fort Ben jm,fn Harrison,<br />

Indiana. The KCS Evelu*tion Test outline ir forwardad to the pertinent<br />

US Amy Schools for their review a.nd revisions. The US Army SChOO16 normally<br />

rely on porronnsl within the School dapartmanto fcr the teat outline<br />

review and revlofonr . After the test outlinas have been reviemd and rcvised<br />

by the US Amy Schoolo, they are returned to the US Army Enlisted<br />

Evaluation Center where a final revision ie performed, and the teat outiine<br />

ia then used by the US Amy Schoolr US a guide when constructing test Items.<br />

When this rycsten it~ ured in reviewing and revising l4CS Evaluation Teat<br />

outlined, it should be noted that ths review and ravleion have been c~ccmplished<br />

only by teat opecialist sod instructor personnel within the Army<br />

Schoolr. For the most part, no anllrted peraonnsl acturlly ?-*orking in the<br />

job hova been involved.<br />

Some of the foulto found in urfng a procedure of thio type for<br />

constructing KOS Evaluatf-n Tert outlines nrt a8 follovs:<br />

a. Publications uaed by test specialists as a baaia in deciding<br />

the aubjcct-matter � :eao to be included fn a teat outline arc in many<br />

carea not current. For exempla, one of the major publications uatd by<br />

toet opeciallst within the US Aczty ‘*en constructing a teat outline ie A job<br />

156


description pbliehed in AR 611-201. This job description lists all the<br />

skills and knowledge pertaining to a portlcular job (<strong>Military</strong> Occupational<br />

Speci al ty) , In many instances these job descriptions have not been upJnted<br />

for several years. There are many other publications relied upon by the<br />

teat specialist chat are net updated.<br />

b. The test specialist and reviewing and rcvistng personnel uho are<br />

constructing a test outline for a partfcular Hflitarj Cccupnticnai Specialty<br />

have had very lfnlted experience with the job in many cases; therefore, the<br />

test outlines might include areas to be tested that are not signiricant or<br />

it might exclude areas that definitely should be sampled. Lack of experience<br />

with the job also results in poo r weighting of the test outline.<br />

c. Some of the terminology used by the constructors of test outlines,<br />

who have had limited experience in the jobs for which the test outlfnea are<br />

being built, is not stated in terms that are understandable to the item<br />

writers and examinees.<br />

The US Army Artillery and MFasile School recognized some of the inherent<br />

weaknesses of reviewing a test outline through a decentralized organization<br />

and began to plan some means of improving test OuKlfne review<br />

prccedures. There was some thou;\rt of sending questionnaires to enlisted<br />

personnel and asking for suggestions on changes to the test outline. Another<br />

idea entertained was that of makin g visits to some Army nrganlrntions,<br />

talking with enlisted personnel,and asking for suggested improvements to<br />

the test outlines. These ideas seemed unfeasible for various r.,asons. It<br />

was finally ihcided that the US Amy Artillery and Missile School would use<br />

what it calls anHOS Evaluation Test Outline Seminar in an attempt to improve<br />

test outline review procedures.<br />

The underlying idea of the HOS Evaluation Test Outline Seminar is to<br />

gamer the thoughts of test specialists, experienced School instructors,<br />

and ffeld experienced anliated personnel at various skill levels and use<br />

theee thoughts when constructing, reviewing, or revising test outlines, thus<br />

creating a test outline balanced on academic school thought, test specialist<br />

thought, and job expertence thought.<br />

These seminars are still !n the experimental stage, To date, two !-lOS<br />

Evaluation Test Outline Seminars have been held. The success of the last<br />

two seminars has indicated that the US Army Artillery and Hissilc School<br />

will use seminars in the future as a regular part of its test outline review<br />

procedures. The last HOS Evaluation Test Outline Seminar was held on<br />

16 September 1964 at the US Army Artillery and Missile School for the pur-<br />

Tose of reviewing and revising the MOS Evaluation Test outline for HOS 142<br />

(Heavy and Very Heavy Field Artfllery Crewman). Those taking part in the<br />

seminar included 20 enlisted personnel, a test specialist from the Enlisted<br />

Evaluation Center, a senior instructor from the Gunnery Department, and test<br />

specialists from the US Army Artillery and Missile School.<br />

157


When a request was made for the names of enlisted peroonnel who would<br />

attend, the request specified that the individuals should be the “best”<br />

qualified in the 142 MOS and have considerable field experience. The request<br />

also indicated that 011 skill levels within the KG should attend.<br />

The active military service represented by the enlisted men who participated<br />

in the seminar rnnged from 10 to 29 years -- a combined total of 296.5 years,<br />

of which 116.5 years are associated with MOS 142. The active military<br />

service per individual averaged 14.8 years. The senior school instructor<br />

who attended had a total of 22 years’ experience in a field cloaaly related<br />

to the MOS. The test specialist attending had considerable experience fn<br />

the field of test construction.<br />

The individuals attending the seminar were thoroughly briefed on the<br />

role they were to play during the seminar and the procedures to be followed.<br />

Personnel were divided into two working groups with all skill levels represented<br />

in each group. The test specialists acted as monitors for the<br />

working groups.<br />

Each member of the working groups was given a copy of the 142 XOS<br />

Evaluation Test outline as proposed by the Enlisted Evaluation Center minus<br />

the weights. The morning session consisted of a review of the test outline<br />

subject-matter area descriptions with all members of the working groups<br />

discc.seing revisions, changes, or additions to the area descriptions,<br />

This allowed the view points of the enlisted personnel, the test specialist,<br />

and the senior school department instructor to be presented on an informal<br />

basis with the consensus determining what should or should not be included<br />

in the test outline subject-matter nreas. The afternoon session was devoted<br />

to the weighting of the test outlil,2 by the working groupe using the aamt)<br />

general procedures aa already discussed. A critique of the day’s work<br />

closed the MOS Evaluation Teat Outlina Seminar.<br />

After the seminar, the test specialists from the US Army Artillery and<br />

Missile School and the Enlisted Evaluation Center reviewed the results of<br />

the suggested deletions, revisions, and changes indicated by the two working<br />

groups during the Seminar and constructed a finalized test outline that<br />

was based on the combined thoughts of the three major groups involved. The<br />

finalized test outline was sent to the pertinent School departments at the<br />

US Army Artfllery and Missile School for their review and comments and then<br />

for-warded to the US Army Enlisted Evaluation Center for final review,<br />

An MCS Evaluatfon Teat outline that has been processed through an HOS<br />

Evaluation Test Outline Seminar results in a test outline that:<br />

a. Contains subject-matter areas significant to the job for which<br />

the examinee is being tested,<br />

158<br />

_-.-----__


. .<br />

b. Contains terminology that should be clear to anyone within the HOS<br />

because the terminology has been vrLtten in terms which has net the satisfaction<br />

of the individuals taktng part in the seminar.<br />

c. Contains properly weighted subject-matter areas because it represents<br />

the combined thoughts of the various types of groups &to were involved<br />

in the seminar.<br />

The main otrength of the HOS Evaluation Test Outline Seminar lies in<br />

its ability to create a test outline that contains a balance of thought<br />

among on-the-job experienced eniisted personnel, experienced school fnstructor<br />

personnel, and test specialists.<br />

The US Army Artillery and Hissile School feels that the HOS Evaluation<br />

Test Outline Seminar will develcp into a valuable tool to be used when<br />

constructing, revising, or relieving test outlines fn the future and it<br />

will provide a vsluable input on which to make recommendations to revise<br />

the NOS job descriptions and other related training publications.<br />

159


Item Writing Procedures for IncrasLng Validity<br />

I. J. WEwEilw, Chairman<br />

6570th Personnel Reeearch Lboratory, US Air Force<br />

The chairman introducad the subject by otatfng that the starting<br />

point for increasing the validity of item8 i6 in the fob analysis and<br />

* the outline. The firrt thing the vrftar must know i8 what he ir eup-<br />

’ I pored to be measuring. If he 10 measuring knowledge, one type of item<br />

.’ 18 required, and if he is mearurtng job akilla, another typa Is called<br />

for.<br />

*,<br />

The particlpantr brought out the vaafour problema they have in thla<br />

area dut to tha vsrfatfonr found in each of the oervlcer approach to<br />

thir problan. When the dircu@eion developed the great interest the<br />

earvices have in a proper job dercripcfon oud job analyaie to help rolvo<br />

thir problem, the choirman reminded the particlpnntc of the work being<br />

done in thfa area by the USAF. He tcuched on thio very briefly,<br />

reminding the group that a fuller triutnent was on the program et a<br />

later reseion.<br />

There aaaned to be agreement that the job analyoie and outYinr<br />

were the foundation of a eood teut item, the problem vipe in getting<br />

this enalysir .<br />

.<br />

160<br />

, .<br />

.<br />

.


\.’<br />

( ,<br />

r<br />

i<br />

I<br />

I<br />

I<br />

I<br />

. .<br />

Non-Fbpfrical Validation of Teat Itms<br />

The folloving paper conctituted n hando\:t. The groop dfncueoed<br />

the pofnte mder la, b, c, and d; nnd 2a. 0, and c on page 4 end 5 of<br />

the peper. No decisive conclusions were darivcd by the group dU@<br />

to<br />

time ?Lmitatione.<br />

The diocussion of the problem preorntsd in this paper shou Id be<br />

fiirther clarbflad, specific e?Xr%QieS furnf.shad, and then ehould ?Je<br />

prarented to the PfYX en u work RmLnar tit the next meting,<br />

161<br />

.;<br />

? !


-<br />

Preoentation to MTA Panel ConaLdering<br />

Non-EmpF;ical Velidatfan of Teet Items<br />

The fact that validity ia many-faced and difficult to determine<br />

does not lessen the importance of the problem. The test produced by<br />

all services here are of tremendous faport. (The results of theoe tears<br />

determine which enlisted personne? get additional pay for isuparior performence,<br />

and help determine which get promoted to the newt higher pay<br />

grade.) Thus, tha ccoaomic and leadership status and quality of the<br />

enlisted personnel of our Armed Forcer are effected etrongly by 0;1r te8tm.<br />

As e result, the morals and quality of our fighting forces ere directly<br />

Affected by the military testing program--our teeto.<br />

It ie our duty to aek ourselves, “How valid are our tertr? Do our<br />

teat identify the best informed and mst nkilled personnel in the euma<br />

compet~t~va occupa~el area? Do thcytruly FdentFfy the least lnfowed<br />

end lee8 rkilled?” Becauee these evaiuatfon testa do affect tha letdership<br />

and morale of the enlleted structure of our nation’s ff.ghttng forcec,<br />

we are interested fn determtnfng and fmprovfng the validity of our itema,<br />

and thus of our testa.<br />

The California Te8t Bureau has cotegorited the two bAsliC spproachrs<br />

to the detcrmlnAtion of validity according to the chr.;t on page 201.<br />

Thorndike end Hsgen (1961) list tha specific considerationa entering<br />

into evaluationa of test8 as (1) validity, (2) rellabFLlty, and (3) practfciality.<br />

Validity refers to the extent to which t test meaeuree what it<br />

18 intended to meascru. Reliability has to do with the accuracy and precision<br />

of a measurement procedure --conoirtency and reproducibility. Practicelity<br />

in concerned with Marty factora, ruch AP economy, convenience,<br />

and interpretabilfty--whfch determine vhether a teixt is practical for<br />

widespread use.<br />

A reading achievement tert requires people to select certain answer@<br />

to quertlonr. Penci.1 marka on enrwer rheeta determlqs each peraon’r<br />

more. Thio Gcore ir called their reading comprehenrlve #core, but the<br />

Icore fs NOT the compreheneion. It io A rfecord of 8ample~behavior.<br />

Any judgment regarding compreheorion ir an inference from thfc number<br />

of allegedly correct anowers. Its vs:fdity in not self-evident but rnuet<br />

be astsblfshed on the basis of BdequAta evidence.<br />

Thorndike and Hagen (1961) #Late, “A teet may be thought of ao cxraapondlng<br />

to some aspect of human bahavior In three sensea.” To identify<br />

these categories they use the terms (1) “reprerent,” (2) “predict,” and<br />

(3) “sfgnify.” Of these three type6 of evidences of validity, our topic<br />

ir prfnci+lly concerned with the fLrst type, KO wit: How well do our<br />

tests reprerment the standard8 (the practical derfred level of performance)<br />

.<br />

162<br />

_ _ .--


--<br />

. .a.,_..<br />

* .<br />

, /<br />

of the occupation or job? Xn proportton a6 the level of job perfornanco<br />

(achievemant of job standards) are accurately represented in the test,<br />

the test is valid. The proceso‘of obtaining thio proportfon of job<br />

performance to test contant is aesentfally a rational or judgmental<br />

one. This 8nalyofa 10 aometLmea spoksn of as rationai or 1ogLcal validity.<br />

Since the analyefs fr largely in terms of the content of the test, the<br />

term contenr validity fs sometimes used.<br />

We should not think of content too narrowly, because we arc Fnterestod<br />

in process as much aa fn atipla content. Thuo, in the test of a<br />

mechanic ve are concerned vl:h euch “content” elements a6 forms for<br />

requioftioning materfal8, of what materials rivets or bolts should be<br />

constructed, rules of safety, priwiples of expanrfon and contraction<br />

and remote reading gages. WC &r-e also interested in such “proceoa”<br />

eklllr as troubleohootFng and correcting defects of equipment, the<br />

ability to solve work aerignment problems, the ability to organize a<br />

repair crew for a particular job, and the use of correct procedurea in<br />

solvfng many other specific job proglanr.<br />

Thus, I submit the therlr that the problan of appraising content<br />

validity of a test starts vith the detailed listfng of the standards of<br />

job performance, and includes rucceo~lvaly ranking the content of the<br />

job standards, and canparing thl..a vith the blueprint of the test. After<br />

thFe task Lu completed, the item mu6t be compared vfth the test outline<br />

and thus vith the job rtandardu.<br />

The extent to vhfch the test outline areas (and thus the ‘CL standards)<br />

arc tapreoan: td in the teat 10 a crueLa fndicator of the validity of the<br />

te6t.<br />

Rovaver , ve cannot ignore predictfve and concurrent valfdity. Before<br />

ve consider techniques or methods of obtatning content validity of tests,<br />

it ir noteworthy to conbider the empirical method of valibting te(ltl),<br />

which Thorndike and Wagen (1961) clasrLfy a6 predfctive validity.<br />

As Thorndike and Hagen (1961) state, and I quote them verbatim, I’...<br />

predlctlve validity can be estimated by determining the correlation betvecn<br />

test scores and a euitablc criterion measure of euccesu on the job. The<br />

joker here ia the phraoe ‘suitable criterion maasure’....One of the moat<br />

difficult problems that the pcruonnel psychologist or educator faces is<br />

that of locating or creating a satisfactory mtaaure of job aucceae to<br />

eervc as � criterion mearure for test valLdation....All criterion measures<br />

are only partial fn that they measure only a part of success on the job,<br />

or only prelfminaries to actual job performances . . .an ultLmate criterion<br />

ir fnaccerrfble to us . ..and substitutes (intermediate criteria) are only<br />

partial and are never completely satisfactory . ..The problems of effectiva<br />

ratfng of paro3nnel arc dfscuasc,! in detafl in Chapter 13. It suffices<br />

to indfcate here that ratings are often unstable and are influenced by<br />

.<br />

163<br />

_ --.---- 3 --


--<br />

--<br />

.<br />

_ . _-._ ._- --- ..--._. - ._.<br />

many factor8 othsr than the proficiency of perronnarl being rated,” The<br />

authors Ilrt mwtn genaral limitations of rating proceduras (pp. 383-384).<br />

Thue, SU& retlnga should salden, if aver, be unod ee the sole or determining<br />

evidence of validity. For example, both logical enalyalo and<br />

raoaarch on ratinga indicate that peara end cuparvisors tend to rnte<br />

eomewhet different qualities end to weight these differently even when<br />

given exactly the same acelco end dafinltlonr.<br />

Whet I em proposing la related to an urea of ccnzmunicatfono rtrcarch<br />

celled “content enelyele.” In his book -Content - Analyaie fn Cmunicmtiona<br />

Rcoesrch, Dr. Bernard Birdson defines content enelyels as?‘...6 reoearch<br />

technique for the objective end quantitative description of the taenifert<br />

content of cmraunfcetlon.” Of courle, the test ltaa le the coazuunlcetion<br />

whooe menlfcet content interest8 ue.<br />

BcfDre proceeding further, perhaps we ten avoid quibbling on<br />

maentlc differencer by agrtclng to UIC English end Englioh (195B)<br />

deflnltlone of certain tat-ma bselc to our dfrcuesion.<br />

Dtflnftlone<br />

.I. Oblectlvt of test: To objectively neeeure the degree to which<br />

a person has mertered eJI1 tlglentr of a epecific occupetion (the vo:d<br />

degree impliea separating workere on the basis of dtffertng ebilftlte.)<br />

2. Job Kaatery: Such pr?fCcleney ln e epecfffc occupation thet<br />

certain definad standards of eccomplfshmsnt can be met perfectly (Englleh<br />

end English, 1958).<br />

3. Discrfmlnetfon: The proceos of detecting dfffcrencte (our teete<br />

must be able to measure differences in degree of job mestary emong pcopl~<br />

working ln the euae occupation).<br />

4. Standard: That which is expected; a practical, desirable level<br />

of ptrformsnce (Englleh and Engllrh, 1958).<br />

5. Job Standard: Army Regulation AR 611-2Gl (1961), pegs 6, atetee<br />

II .** d. Fzth digit . . ..Thara aklll level designations indfcate the level<br />

of proficiency required in e epeclflc job end the corresponding quellficetlon<br />

of an lndivlduel... (With the exception of HGS Code 718) NAVPERS<br />

18068 (1958) lieto the job � tendardo of Navy enlleted occupation groupr.<br />

Air Force Kenuel 35-l (1957) lleto the job standard8 for the Air Force<br />

enlirted occupatione. Slmtlar regulations list job etendardr for the Royal<br />

Cenadien Navy, the US Coert Cuerd, the Herchent Heri.ne Academy, and the<br />

tfarlne Corpr .<br />

.<br />

.<br />

164<br />

--


,<br />

‘ .<br />

. .<br />

_ -. . .<br />

For other deffnitlon&, English and Englierh (195Y), fn available<br />

for reference if neeP be during thir diecuosion.<br />

A teat itm la only valfd to the degrees that it represcntrc the<br />

job standard it la meant to rample and to thfa dqrea it contributes<br />

to the validity of the teat aa a whole.<br />

In order to guide ua In makin g useful prozrsoo Fn the ohort time<br />

available, I have llated major rteps which should help UQI to work orlt<br />

an acceptable proceao for logical validation of teat outlinen and tent<br />

items.<br />

1. For logical validation of test outlince vc muat accomp’iiah<br />

the following:<br />

a. Datamine what conrtitutto the panel of exports.<br />

b. Determine the criteria which the panel of experta plhould<br />

use in ranking element8 of job standards.<br />

c, Detel-mina the methode to usa in correlating the test outline<br />

with its aealgned weights -a8 it -axiete - with the rank order of job<br />

standardr arrived at by the panel of experto.<br />

d. What criteria, if any, ohould propored changee to teat<br />

outliner be required to meet before they are incorporated into cxirtfnp,<br />

validated teet outliner?<br />

2. For logical validation of teat Itemme, we uuat accomptfrh the<br />

following:<br />

a. Detennfne what conrtftuter a panel of experts. (Should<br />

they be the same peroonnel who validate job rtdndards?)<br />

b. Determine the criteria vhich the panel of experts ehould<br />

use in ranking teat Itema.<br />

c. Determine the methods to use in correlating teat ftmn<br />

vith the following:<br />

(1) Original tc6t outline 4rerc.<br />

(2) Rank order cf fob standards (tert outline) obtained<br />

fran the panel of experts.<br />

165<br />

____ _________ _ __.. --.___ ,_.___. __--- --<br />

I<br />

I<br />

I .


.-<br />

We probably ohould not conc1~rn ourcralvea here ~Lth tha tcchnlcnl<br />

accuracy of itakka in 43 test, EB rhie point rel.acwi to content validity<br />

uauafly fe validated by chockFug tha publication uaeci Fn raforencing<br />

the LeQm.<br />

A point which ahould bn coneidarsd ia that such rtrmrch BI has<br />

been conducLed hm failad to ffnd my eigntffcmt diffarauce In validity<br />

of ftcme ragardllaeo of uhothsr they conforra precFesPy to “Ltesn construction<br />

prfncfplar” or not.<br />

1. Fraquency of perfaming tho task: Vow hportant ie thio<br />

factor? What is tha Polorionehip of frequency, or routina, of job took<br />

to ehe crdtfcalit~ of ths tark?<br />

2. Criticality of performing the task: What fa the rerult if<br />

the tank ir HOT parfowed? if the task io ?JOT perfowod correctly?<br />

rafaly? Will man be killed? aquipmant ruined? miasfon fafl to be accompliohad?,<br />

etc.<br />

3. KncwZedgs ssaential to perform the feek.<br />

4. PTOC~YI, or use, of the knowladge to perform eha raako,<br />

5. Hew well do the problems in thirr ~1rm discriminate between<br />

anlisted perrono having high, svaregc,, and low ?svaln of job curtaty?<br />

Z’ha ocalar attached ara rugeasted for your use in l.le~lng auccfxtIy<br />

tha oseential criteria axparr judgea ohould u5?1 in cetegoriting job treks<br />

and teat itema atz one and of ehe scalao or in the middle. Scratch ; aper<br />

10 ateachod aloo for your coiweniance.<br />

E: Savaral copieo of the content validation etudy being conducted<br />

by Major David Culclasurs, MS Evaluation Teat Project Director, US Arq<br />

Medical Field Service School, Fort Smn Houston, TBXEO, ara availatlo to<br />

panel mexnbare for information and uce for suggaotloar do to procedurar.<br />

We invite Major Culclaaure to ditcues hia study with uil at thle point.<br />

166<br />

.


.<br />

-----<br />

We probably ahould not conc~~tn ourselvGa here with tho technical<br />

accuracy of item8 fn a test , us this pofnt ra?ot%d to content validity<br />

urually ir validated by checking :ha publication used In raferenclrq<br />

the It=.<br />

A point which ohould be considered ir that much raeearch aa hee<br />

bsen conducted harr foilad to fIn.3 sny oignificant difference In validity<br />

of ittwo ragardleor of whether they conform pracfaaly to “ita construction<br />

prfnciplae” or not,<br />

1. Frequency of parformIng the task: Ho3 inpsrtmt is thio<br />

factor? Whet Ia the relationehip of frequency, OT rcuclne, of job ttask<br />

to the criticality of the taok?<br />

2. Criticality of perfoming the task: What Is the rasolt if<br />

the Lark Ir NOT performed? If the task IO NOT parfamtd corrsctlyt<br />

� afaly? Wfll men be kIllad? squipmant ruined? miosion fail to ba accomplishcd?,<br />

etc.<br />

3. Knowladga eratntlal to parfom tha tack.<br />

4. Proceea, or UIQ, of the knowledga to perfom the taoks.<br />

5. How wall do the problems In this ares dIocrLmZnate between<br />

-lIlted pereons having high, average, and low levalrr of fob rnnetery?<br />

The rcnlar attached ara suggested for your uoa In listing ruccfnctly<br />

the essential criteria oxpert jud305 should use in categorialng job tmrko<br />

and toet Items at onr end of tha scales or In tha niddle. Scratch pager<br />

I8 attached alro for your convenience,<br />

NOTE : Several copier of the content valIdrt?an otudy being conducted<br />

by MaFDovtd Culclot~rti, MOS Ev&luatIon Test Project Director, US Amy<br />

Medical PIeld Service School, Fort Smn Rouoton, TEXAS, arc avaflatle to<br />

penal Iombarr for Infomatton and use for suggertions aa to procadureo.<br />

We fnvfte Hajor Culclavure to diecuos hia etudy with ua at thic point.<br />

166<br />

,<br />

c


. .<br />

.<br />

LOXCAL EKPIRICAL<br />

Content Valldfty<br />

- -<br />

Refers to how vcll the content of<br />

the tent samples the subject matter or<br />

� ituatioo about which conclusions are to<br />

ba dravn. Content Valid& is especially<br />

important Ln an schicvenent teat. Rxamplea<br />

: testbook analysis, description<br />

of the universe of itena. rdoquacy of<br />

the eI[Icple, represanta~:ivencss of the<br />

teat content, intarcorr?lationa of<br />

subocorcs, opinions of jury of experts.<br />

1tcm structure<br />

Includes (1) Corroborative evidence<br />

from item analysis rupporting the other<br />

characteristics of the teat; i.e.. interrclntionahips<br />

between items, bctvecn items<br />

and scores, and between items and criteria;<br />

(2) Item ccmposftion. For graphic items,<br />

it emphasizes perceptual clarity and rclated<br />

format functions. For verbal itms,<br />

i< mphasircs conceptual clarity in the<br />

cqra8oion of itema. For both graphic and<br />

verbal items), it Emphasizes functions of<br />

dietractors.<br />

I<br />

Construct Validtty<br />

I<br />

Concarns the psychological qt.-lities e teat<br />

mcaaurco. By both logical md empirical mcthodr<br />

the theory und’erlying the test is validated.<br />

Exmpleo: Corrciatims o f the teat scorea, factor<br />

malysir, uoc of inventories. studying rhc effect<br />

of speed on test scores.<br />

I<br />

A<br />

<<br />

. $, v<br />

-Cr-..-..~..--._-,---<br />

Predictive Validity<br />

Relates to hw veil predictions u.ade<br />

rfrom the test are confitued by data collactcd<br />

at a later time. -loo‘? Correlations<br />

of intelligence test8coras with<br />

cwrfe gradeo, corrclatioo of teat acores<br />

obtained at beginning of year with msrka<br />

earned at the md of the yaLr.<br />

Concurrent Vafidlty<br />

Refcra to hov veil teat mores natch<br />

mcasuree of contcmporsry criterion porfomance.<br />

Examplea: Cowparing of scores<br />

for men in an occupation vith these for<br />

men-in-general, correlation of pcraonaPity<br />

test scores with eotiwtcs of adjustments<br />

made in the mmnrcling interviews, correlation<br />

of end-of-course achieveac:rt or<br />

ability test scores with school okra.<br />

.<br />

\<br />

___~__^<br />

1<br />

I<br />

,,* - _,. . - - * -.<br />

.<br />

.


I<br />

Y--L.<br />

.<br />

.-<br />

-. *<br />

-- ,<<br />

.<br />

_.<br />

:<br />

.<br />

-<br />

: ,-<br />

\..L<br />

. -<br />

-=Yp<br />

,<br />

. .<br />

._ -*<br />

- .-<br />

L<br />

.C<br />

-._<br />

.


,<br />

__-____ 14_1- -...-. -- _.._.~<br />

_..<br />

..___ -..~ .- _.._ -___-<br />

___C__-..--<br />

169


,<br />

c<br />

References<br />

Birdaon, B. Content analyeir in coimunication~ reaearch.<br />

- - -<br />

Illfnoff3: Free Pfesn, 195X<br />

Glencoe,<br />

Californfa Teat Bureau. p- gloeeary of meaeumment tax-ma. Monterey,<br />

California.<br />

Chief, Bureau of Nuvol Peroonnel, US Navy, NAVPERS 18061. Washington,<br />

D. C., 1958.<br />

English, H. B., b English, A. C. Dictionary of psychological and<br />

psychoanalytical terms. New York: LongGiG & Company, 193K<br />

Headquarters, Department of the Air Force. Air Force Manual 35-l.. 1957.<br />

Headquarters, Department of the At-my. AR 611-201, Washington, D. C., 1961.<br />

horndike, K. L. & Hagen, Elfrabtth. Meaaurcnant and evaluation in<br />

psychology and education.<br />

Sons, Inc. ,196l.<br />

(2d ed.). New YorrJohn Wiley &-<br />

,<br />

170<br />

.


,<br />

.- -<br />

L<br />

L<br />

.<br />

-.<br />

.--Ly<br />

-_<br />

_. -<br />

. .<br />

\.<br />

\<br />

. .<br />

. j<br />

‘r<br />

:\<br />

\ I ;<br />

.<br />

. .<br />

Teet and Item Revision Technique8<br />

J. E . PARTTHCXOH , Chairmsn<br />

Teat And item rovisdom Are neceenuxy for A vcnriety of ressone.<br />

‘Zrc main reason for euch revisione ie to bring About improvement in<br />

measuring inotrumants. A test muat be revised if the job requiramcnte<br />

are changed And if it la not functioning properly AA A mcssurfng inatrumcnt,<br />

Among the pany raaAonA for reviaing itema Arc: obsoleaonca, too<br />

easy, too difficult, do not discriminate bctveen those who have HIsatered<br />

the job and those who have not, lack of validity, not job oriented,<br />

require rote memory to ancrwar.<br />

Techniques for Item Revision<br />

Statistical data, particularly the item anelyaio capl Aerve ca a tool<br />

or guide for indicating area8 vithin iteazre which Ara in need of revieion.<br />

It ie neceesAry for the teat psychologiet And the subject-matter expert<br />

to work together in analytfng And literally “taking apArt” the test item<br />

which According to tha statieticel Analyale hea not functioned in the<br />

way that it rhould. Evan though tha subject-matter expert May knc~a hio<br />

Aubjact,ha rosy not be Able to analyze the poorly-functioning item And<br />

bring about changes in it. it iA usually helpful for the teAt peychologist<br />

to ark the eubject-mattsr expert to deacriba in detail the protees<br />

or actions through vhfch tha eraminae must go to Bnevor tho particular<br />

Item. This should bring into clear focus for tha rubjact-mattcr<br />

axpart tha Apacific raquiratants on the part of tha exeminaa vhan he<br />

is facad with A problem which the Item presenta. When thio Approach<br />

iA tAkan to item revieion, it is oeu~lly a~sy for the eubjectwttar<br />

axpart to caA ths reAAon for tha poor functioning of the item.<br />

The following Actione or consideration of the following fsctorr<br />

rhould ba helpful in the revieion of test itatnr.<br />

A. ItemA rhould ba 80 written AA to make thafr dirtroctorr AA<br />

AttrACtiVa AO poasibla. . .<br />

b. Look for undatccted Ambiguity in fttxn diotractoro which<br />

mirlsad sxcrPrFnaea. Itaa Atems May also be ~mbiguour.<br />

c. Eliminate nonfunctioning dietractors.<br />

d. UIQ hints provided by the dAtA concorning mentA1 procarrso<br />

of axzunlnaes.<br />

I ---.---_.----___ ._ . . --. ---__<br />

171


‘7-.<br />

* -.<br />

,<br />

.<br />

--<br />

.<br />

_ ._<br />

e. If a diatractor ia discriminating in the wrong direction it<br />

may rapreoent an aspect th8.t cannot be r8movod without d88troying the<br />

whole point of the item. The iccm nny be covering the point about which<br />

there fo much mietnformntion; the dintractor should not be revlaed if<br />

the point of the item ie loat through ravieion.<br />

The following quotation from Devie’ ch8ptar on “Item Selection<br />

Techniques” in Lindquist’e Educational Measurement should be helpful<br />

when revieion technique8 8re applied to testa end item8. .-<br />

“It is interesting that mny invalid dietracters are found in items<br />

that how been carefully edited and checked by subject-matter urperts.<br />

Thi8 anph88ixes the well-kno-m fact thnt because a dfatracter ie dircriminative<br />

in the wrong direction we cannot conclude that it io too<br />

nearly correct from a factual point of view. Convereely, the fact that<br />

an incorrect choice ie too nearly a correct anawer doe8 not nacasaarily<br />

mean that it will turn out to be diocriminative in the wrong diraction<br />

when it ir subjected to item mnlyoie. Item analyois technique8 cunnot<br />

alone be relied upon to detect error8 and 8mbiguitieo; expert crfticiw<br />

and editing ore indiepenaable in te8t construction, The full value of<br />

item analyrie technique8 cannot be reelircd unlers criticiomo of the<br />

item8 by recognited authoritier are available for reference.”<br />

Referance<br />

Davio, P. B., Item selection technique8 in Lfndquiot, E. F. (ed.),<br />

Educational Measurement. Uarhlngton, D. C.: American Council<br />

on Educcltionx951, 308.<br />

.<br />

.<br />

\<br />

,<br />

172


._“...-<br />

.<br />

,<br />

._ I . . ..,_ . _, _ . . .-<br />

.<br />

__ .I ^., bL<br />

PAPEA<br />

Answers to Common Criticisms of Tests<br />

FRANK H. PR-ICE, JR.<br />

US Army Enlisted Evaluation Center<br />

It is appropriate at s conference of this type that we constder sane<br />

of the common criticisms of tests. This consideration is especially fitting<br />

since the mass testing movement was spawned as a result of the rush<br />

to military prcperdncss for World War I.<br />

The criticisms of testing come from many sources both within and<br />

without professional psychology. Generally, the psychologist critics are<br />

constructive while the lay critics are destructive. This morning we are<br />

concerned with these lay critics who have protested in the popular press<br />

and books such as Hoffmann’s The Tryanny of <strong>Testing</strong>, Harrington’s Life in<br />

the Crystnl Palace, Gross’s The Brain Watches, and Whyte’s Crgnnfzation<br />

&, indicating public distrust, uneasiness, and ignorance about which<br />

we mus’t be concerned,<br />

Most of our critics have leveled their blncts at so c6lled educational<br />

and industrial tests. Very seldom have our military testing programs been<br />

the direct victims cr’ such scath:ng attacks; but, just because we have not<br />

been the subjects a; eloquently set forth pronouncements does not mean<br />

that we are not discussed and “cussed” in the dayrooms and barracks of<br />

those we test and even in the offices and headquarters of those for whom<br />

we test.<br />

My remarks this morning will not be limited to testing in the military<br />

setting. Almost all of the recent protests against psychological<br />

testing are as applicable, if not more so, to the testing of military<br />

personnel as to the testing of school children, college students, or<br />

Industrial applicants and employees. And I hope that you will be able<br />

to apply this discussion of some of the criticisms in terms of the<br />

particular problems and Interests of your military programs.<br />

In this vein, I will devote particular attention to The Tyranny<br />

<strong>Testing</strong> (1962) primarily because Hoffmann appears to be more sophisticated<br />

atld-devious, but slightly less venemous than other of our critfcs. To<br />

give credit where credit is due, a symposium on this subject presented<br />

by Owens, Astin, Dunnette, and Albright at the 1963 meetings 4 t’,e Midwestern<br />

Psychological <strong>Association</strong> has provided much of the source material<br />

for this paper.<br />

First, let us examine some of the major assumptions found in<br />

Hoffmann’s detailed indictment of the multiple-choice test. He (p. 150)<br />

states, and here he is talking about multiple-choice tests -- the same<br />

type we construct and admfnfster -- “The tests deny the creative person<br />

a significant opportunity to.demonstrate his creativity and favor the<br />

shrewd and facile candidate over the one who has sometning to say.” The<br />

173<br />

, I<br />

_ . _ . _


.-<br />

/ .‘4 ‘._.<br />

.<br />

- --<br />

.<br />

problem with such an ammption is that there lo virtually no sotieftctory<br />

research defining creativeness. Usually the people who sccre high. on so<br />

called “creativityI’ tests are aimply the ones we call creative. In fact,<br />

there is considerable evidence to indicate that creativity teal& arc not<br />

actually measuring creativity as a personality trait (Thorndike, 1963).<br />

There is no evidence from carefully conducted and logically interpreted<br />

research to indicate that objective tests stifle the creative person.<br />

Roffmsnn’s charge is what he thinks should be fact rather than research<br />

data. In other words, Roffmann has the idea that merely because multiplechoice<br />

tests are highly structured, the examinee has nc opportunity tc<br />

expr.?ss himself. Nothing could be further from the truth, but the degree<br />

to which the examinee can express his knowledge depends on the skill and<br />

the data of the test writer.<br />

In his second assumption, Roffmann states that multiple choice tests,<br />

,I . ..penalize the candidate who perceives subtle points unnoticed by less<br />

able people including the test makers. They are apt to be superficial<br />

and intellectually dishonest with questions made ortifically difficult<br />

by means of ambiguity becnuse genuinely searching questions did not<br />

readily fit into the multiple choice format.” In this assmption the<br />

grent amount of careful research actually going into the construction<br />

and validation of a teat item is completely ignored. Naturally, dierractors<br />

are written purposely to “fool” the less knovledgoa’ble examinee.<br />

information about the responses to items m*de by persons of different<br />

levels of knowledge indicates without n doubt that the degree of ambiguity<br />

perceived by an examinee is inversely related to his knowLedge of the<br />

subject matter, This simply menns that the less one knows the more<br />

ambiguous the question appears. Yet, Hoffmann state6 (p. 67), The more<br />

one knows about the subject the uarc glaring the ambiguities become.”<br />

Of course Roffmann doea not support his assumption with evidence;<br />

nevertheless, this charge is the one with which we are moat often hit.<br />

He saye that the most serious consequence of test ambiguity is that it<br />

penalizes the gifted and talented examinee. In Hoffmann’ view, how does<br />

this discrimination occur? When first confronted vith the alternative<br />

answera to a question, the “deep” examinee, as Hoffmann calls the gifted<br />

and talented, analyzes the alternatives more carefully than does the<br />

*‘sup’ -ficial” exsminee. Naturally, such careful scrutinizing takes tfme,<br />

and the first penalty occurs.<br />

Secondly, the “deep” student is much more<br />

likely to perceive the ambiguities and, as a result, spends more time<br />

trying to determine exactly what the test author had in mind. Furthermore,<br />

according to Roffmann, the “deep” examinee’s motlvetion to perform<br />

well tends to be reduced as he sees more clearly the superficiality and<br />

ineptness of the test writer’s approach. Even more damaging, the gifted<br />

examinee is more likely to discover a “better” al.ternative than the keyed<br />

response.<br />

174<br />

.


c r-<br />

’<br />

,<br />

L-<br />

.’ ./’<br />

/ :‘..<br />

problem with such an assumption is that there i.8 virtually no satiafsctory<br />

research defining creativeness. Usually the peopie GAO Gcore high on co<br />

call.ed “creativity” test6 arc simply the one0 we call creative. In fact,<br />

there is considerable evidence to indicate that creativity testd are not<br />

actually measuring creativity as a pereonality trait (Thorndike, 1463).<br />

There is no evidence from carefully conducted and logically interpreted<br />

research to indicate that objective teats stifle the creative person.<br />

Hoffmann’s charge ie what he thinks should be fact rather than research<br />

data, In other words, Eoffnann ha8 the idea that merely because multiplechoice<br />

tests are highly structured, the exminee has no opportunity to<br />

expr.?ss himself. Nothing could be further fros the truth, but the degree<br />

to h%ich the examinee ten express his knowledge depends on the skill and<br />

the data of the test writer.<br />

In his second assumption, Hoffmann states that multiple choice tests,<br />

1, . ..penalize the candidate uho perceives subtle points unnoticed by less<br />

able people including the test makers. They are apt to be superficial<br />

and intellectually dishonest with questions made artifically difficult<br />

by means of ambiguity because genuinely searching questions did not<br />

readily fit into the rmrltiple choice format.” In this assumption the<br />

great amount of careful research actually going into the construction<br />

and validation of a teet item is cmpletely ignored. Natur.tlly, distr8ctor6<br />

are written purposely to “fool” the less knowledgeable examil.se.<br />

Information about the responses to iterns made by persons of different<br />

levels of knowledge indicates wlthout a doubt that the degree of ambiguity<br />

perceived by an exuainee is inversely related to his knowledge of the<br />

subject matter. ‘fiftl simply means that the less one knows the more<br />

ambiguoue the question appears. Yet, HoffmaM states (p. 67), “The more<br />

one knows about the subject the rare glaring the a;tilguitice become.”<br />

Qf course Hoffmann dccs not support hie assumption vith evidence;<br />

nevertheless, this charge is the one wit.h which we are most often hit.<br />

He says that the most serious consequence of test ambiguity is that it<br />

penalizes the gifted and talented examinee. In Hoffmann’s view, how does<br />

this discrimination occur? When first confronted with the alternative<br />

answers to a question, the “deep” examinee, as Hoffrrann calls the gifted<br />

and talented, analyzes the alternatives more carefu2ly than does the<br />

l’superficialt’ exnminee. Naturally, such careful scrutinizing takes time,<br />

and :-Be first penalty occurs, Secondly, the “deep” student is much more<br />

likely to perceive the rrmbiguitlee and, as a result, spends more time<br />

trying to determine exactly what the teat author had in mind. Furthermore,<br />

according to Hoffmann, the “deep” examince’s motivation to perform<br />

well tends to be reduced as he sees more clearly the superficiality and<br />

ineptness of the test writer’s approach. Even nore damaging, the gifted<br />

examinee is more likely to discover a “better” alternative than the keyed<br />

reeponse.<br />

%<br />

174<br />

.<br />

-____-- - _.<br />

.<br />

.<br />

.


f<br />

Ih<br />

i<br />

i<br />

Hoffmann attempts to document his reasoning -- and it in mere11<br />

U’arm-chairing” -- that multipie-choice test qucctitions are by neture<br />

ambiguous by citing Pample items from test manuals or more frequently<br />

by attacking illustrative items of his 0.m. A Hoffmann-type item C.W<br />

quickly illustrate the kind of rumination which forms the main content<br />

of his attack,<br />

i<br />

.<br />

I .<br />

Watt are the colors of the American flag?”<br />

(A) red, white, and blue<br />

i<br />

c<br />

, 0)<br />

((3<br />

say<br />

neither A nor B<br />

i The superficial examinee quickly selects answer “A,” red, white, and blue,<br />

and goes to the next question. The “deep” examinee begins to scrutinize<br />

and analyze the alternatives; he thinks, “A” is correct under some<br />

conditions, but “B,” gray, Is correct under some conditions too -- twilight,<br />

poor illumination, total color blindness. Both “A” and “B” could be correct<br />

under some conditions, but the other alternative, “neither A nor B,” could<br />

not be correct. Supposedly, ha wonders if the questfon is a trick, if<br />

the test wrftcr is malicious or just plain ignorant, and what was really<br />

wanted. Finally in desperation, t,e throvs up his hands and says, llThie<br />

is an absurd test nnd 1 don’t see how any intelligent person can be asked<br />

to take it seriouslylf’ And 1 have heard just that in the field on more<br />

than one occassion.<br />

This charge of ambiguity is one of most serious consequence to us<br />

and one on which we are most vunerable unless we consider the fundamental<br />

distinctions among the purposes of test and test items. I am referring<br />

to the distinction between tests which are used as criteria ant. tests<br />

which are u6eu as predictors. - -<br />

Criterion tests -- for example, achievement<br />

tests for comparing the effects of dffferent methods of teaching -may<br />

be open to Hoffmann’s criticisms. But his objections are lrreievant<br />

in the case of predictive tests -- for example, tests to select the most<br />

promising job applicant or the best qualified soldier for a special assignment.<br />

Basically a criterion test must be content valid -- that is,<br />

we must be able to defend the test on rational and logicalgrounds. All<br />

that fs required of predictive tests is that they sucessfully predict<br />

future performance.<br />

Now where does this leave those of us who use our tests for both<br />

purposes? Naturally we must attempt to demonstrate that vdt tests have<br />

both content validity and predictive or concurrent validity. It appears<br />

chet Hoffmann has never heard of empirical validation, but he manages<br />

to marshal1 a variety of defense against the evidence. He particularly<br />

attacks the criterion sgainst which we validate our tests. However) the<br />

value of a criterion is a very different question frocn the problem of<br />

whether a test can predict that criterion. Criticizing the tests simply<br />

175<br />

t


ecause we do not like the criteria is confusing the issue. Hoffmann’s<br />

concern rfth ambiguity is relevant only when the content vaLidity of a<br />

test is the primary constd.eratFon. Eve,1 80, tke customary item acreesing<br />

and selection in the constructicn of teats insurea a reaeonably satiafactory<br />

degree of content valfdity. This io particularly true in the<br />

case of our military tests which are designed to cover specific job areas.<br />

What about Hoffa%M'S cLaim that the “deep” examinee fs penalized<br />

by multipLe-choice iteM? What he seems to be saying is that the very<br />

dull examinee will fail the itea more frequently than the superficially<br />

bright examinee, but that the exceptLonally bright examinec vi11 have<br />

more trouble with the same item than the superficially bright examinee,<br />

Naturally, the brightest examinees are more apt to discover ambiguities<br />

than are other students. But Hoffmann’s ideas about the consequences of<br />

their perceptiveness do not hold up.<br />

To get back to Hoffmann’s assumptions, he states that tests,“... take<br />

account of only the choice of answer and not of quality of thought that<br />

lead to the choice,” and “They neglect skill and dtsciplfned expression.”<br />

This attribute he catls “quality of thought I’ is not defined, nor is any<br />

reliable and valid way of measuring it suggested. One mfght assume that<br />

Hoffmann would advocate use of the essay examination as a measure of his<br />

quality of thought. Such is not the case, for he outlines no less than<br />

four very convincing arguments why the essay test should not be used for<br />

that purpose. (L. Difficult to choose a topic fair to a% 2. difficult<br />

to determine whether the essay is actually revelant to the question; 3.<br />

difficult to overcome the problem of negatfvc halo due to poor handwriting,<br />

spelling, etc. ; 4. difficulty to maintain consistency within and between<br />

scores.) I often hear this quality of thought argument twisted around to<br />

say that just because a man scores high on the test does not necessarily<br />

mean that he can do the job. Most generaLly, those advancing this<br />

argument mean that the man will not do the job. While it may be true that<br />

a high score on an achievement or’ job knowledge type test does not<br />

automatically indicate a high degree of motivstion, it is eqtially true<br />

that the Low scorer cannot perform the fob, that is, he does not have<br />

the knowledge to perform regardless of hie motivation. Of course this<br />

depends on test validity and test purpose. It is up to us to define<br />

the purposes to which oar instruments may be put and to determine the<br />

validity, for various purposea, of these instruments.<br />

At this point we might conclude that Hoffmann does not like any kind<br />

of testing. We would probably be correct. He criticizes objective tests;<br />

yet he leaves no alternatives for mensurement. When he argues that &he<br />

results of an educational testing service st*)qy showing an essay teat<br />

was less good than an objective examinstion are silly and could not<br />

possibly have been obtained, he is simply confusing content with predfctive<br />

or concurrent validity.<br />

176<br />

,. ^ .._ -.. .__.-<br />

.


.<br />

. . . _. . - ._<br />

.<br />

. .<br />

Perhaps tha potentially mo6t damaging a66w$ion, rind the one which<br />

would be the mont difffcvlt for R0ffraan.n to 6u6tain, has to do with the<br />

effect of teatrr in the identification of individual mertt, He 6tate6,<br />

“They have a pernicious effect on education, and the recognition of merit.”<br />

Thi6 io cf cour6e -wholly without fcwundation. There have been too many<br />

6ucce66 6torie6 for u8 to even bother to refute hio claim.<br />

Now let u6 briefly turn to 6ome of our other critics. A major as-<br />

6umption found throughout the =fCfnpe of Packed, Barrum, Grose, tatlFte,<br />

Xarringtm, ard other6 i6 that “mind” cannot be meaeured. Tbcy seem to<br />

Consider ‘hind” a human “mystery 6y6tem” outeide the realm of rcientific<br />

otudy. This may be true, but if “Axi” is defined aa behavior, it indeed<br />

can be measured. In fact, the measurement of man’s fndividual differences<br />

perhapr ha6 been the greate6t sccwplishment of psychology thus far. We<br />

can anaeas the individuality of persons rind make pretty good predicitions<br />

about their future behevior.<br />

Our critics make tin assumption exactly opposite to the aerrtqtion<br />

I just mentioned. They se+- that test leads to conformity by picking<br />

perecmo who are all of the same type. This, of course, as6ume6 that<br />

“mind” can be measured only to well -- in fact, considerably better than<br />

we are able to measure it. Anyvay, there is considerable evidence to<br />

substantiate intra-individual trait variability. Thfa principle of<br />

p6ychologlcal testing 16 fgnored by these critics, The feet that people<br />

do differ vithfn themselves has long been recognized srnd submitted to<br />

careful etudy by peychologfste.<br />

It i6 complained that testing is an inva< ?on of prlvocy, and this<br />

criticicm m.6~ have 6ai~e merit. It i6 incumbent upon ‘:6 to demonstrate<br />

the validity of any item6 which might otherxiae be regarded a6 such an<br />

invarion. In the military eetting we arc seldom bothered by rhfs problem,<br />

especially in proficiency evaluation.<br />

The critic6 very rarely euggert alternative8 to p6yChOlOgiCal testing.<br />

Gardner, in hi6 book, Excellence, 6aye,“Anyone attacking the ueefulnerrr<br />

of te6t6 6w6t suggeet workable elternatlves. It ha6 been proven over and<br />

over again that the alternative method6 of evaluating ability arc subject<br />

to groso error6 and capable of producing grave injucticcr.” WC have only<br />

to look et our = rating ryrtem6 to realize how truly Gardaer rpoke.<br />

Or we could return to the day6 when perronnel decisions vere made on the<br />

baeir of hair color, family background, shape of the head, or sate other<br />

equally intuitive basis. The alternatives to testing when they are euggeeted,<br />

are cIearly rldfculous.<br />

So far thir morning,1 have attempted to point out the merits of our<br />

crltlc’6 charge6 agafnat testing, and we can conclude that the proterts<br />

tith oubatance are those about which we already knew. Hov we ahould take<br />

a look at the impact of our crftica. I do not believe that the Mllttary


--<br />

.,<br />

-__. . . .<br />

,Establlshment has become disillusioned with testing any more than I<br />

believe the general public. has. This conference is evidence that the<br />

Armed Forces are among the strongest supporters of testing. T!l2 protestors<br />

have not caused any detectable eifect in terms of reduced sale8<br />

of testfi or testing ecrvicee, The number of letters from educator6 or<br />

the general public (to reputable test publishers and the APA) critfciztcg<br />

teats has not increased. Generally, the poFuler pre6e hae not j-pad<br />

on the band wagon to foment a public outcry against u3. Apparently, the<br />

net effect has been little more than a few booka sold and a little high<br />

blood preesure BJXXXI~ psychologists. The latter at least may not be such<br />

a bad thing. Some of us need to be aroused.<br />

If this is the case, then why should we even bother to consider<br />

these self-styled protectors of a tyrannized society? The writings of<br />

these critics should make us stop, look, and listen. If we are to avoid<br />

future trouble and Improve the state of our nrt and aclence. we must<br />

improve the quality of our tests and services. He uwst adhere more<br />

etrictly to the ethical standards of our profession. We must give<br />

greater attention to the technfcal cha:acterfstice of our teats and<br />

criteria. We must continue our ef:orts to expose snd elimfnate the<br />

quacks and incompetenta who dwell about the fringe of psychological testing.<br />

And we muat improve our communications about testing 4th every<br />

segment of our public which we can reach. These cmdnicutions met be<br />

technically sound but they also must be written in underatandabl.. English.<br />

We might even form a committee withfn this association to draw up a pmphlet<br />

of general testing principles. and practices &ich could be diesemfnated<br />

within the <strong>Military</strong> Establishment.<br />

In conclusion, the basic assumptions of our critics are crroneoue<br />

and ftIliscfou8; they are generally baaed on lack of information, aa<br />

apparently is the case of Hoffmann, or, more seriously, on a refusal to<br />

accept the strong eolpirical evidence showing that fudfvfdurlity can be<br />

accurately assessed in such a way as to give better recognition toreal<br />

merit than has ever before been the case in our educstional, fnduetrinl,<br />

or rillitory lnetitution5.<br />

Actually, we know that standardized objective testing is ona of<br />

the great success stories of our day. This has been no better pointed<br />

out than by Gardner, in his book, Excellence (1961). Psychological testing<br />

for the firat tine enables us to look at the many facets of an individual<br />

rether than making judgmenta based on the oo-called “lump of dough”<br />

doctrine. Now we can truly measure and assess the individuality of each<br />

of our military personnel and through careful guidance help each individual<br />

realize hts potentialities ae indicated by our peychological<br />

.<br />

.<br />

178<br />

--_


\’ .<br />

c<br />

.<br />

,<br />

.<br />

* .<br />

. .<br />

testing iistmente. Tests provide us with the best u~ana avnilable for<br />

osseseimg individual profi.ciency and discovering md rewardfng indiv?dudl<br />

merit. We can help the cmadcr i:n t’he f!.eld indfviCuaII.y and different-,<br />

iaT!.y carei’ul.ly exnoine his men so ther he no longer mJSK depend cn his<br />

“feel-” or geatoult for pernonnei declaions, This ia cur gresteot strecgth.<br />

When the facts are Iaid it UnderstandabLy, it cannot be disputed or<br />

refuted.<br />

179<br />

-- -. ._^__ --._ _-... __ . - _. -. - _ ____---<br />

I


.<br />

6. .<br />

. .<br />

.i 4<br />

*<br />

.I<br />

.J ,I<br />

--.<br />

Dunnotte, H. D.<br />

prediction,<br />

Reforencrs<br />

Some methods for enhnncLng the vsltdita of prJycholoam<br />

PG read at ~%ympo~i~.~ ~Subgtoupin~Analyois a& AXI<br />

Approach to the Predfctton of lindfvidual Behavior, St. Lwia. MO,, 1962.<br />

Ethical otanderds of pcychologfete. Amer. Psychologist, 1963, 2, 56-60.<br />

Cdrdnor, Y. W. Excellence. New York:Horpar, 1961.<br />

Hoffman, B. The tyranny of teetine. New York:Crowell-Collier, 1962.<br />

X80<br />

I<br />

--1-- --...--w<br />

--


.<br />

Summary <strong>Report</strong> of the Steering Camnfttee<br />

The 1964 Steering Committee diacussfono prodklced the following<br />

resulto:<br />

a. The preaent mA emble3n Wit8 considered inadequate. After<br />

viewing a number of sump18 emblems, e bnafc deeign vas selected to be<br />

modified according to certain specificatr>na. It was agreed that the<br />

emblem would be drawn up and submitted to the Servfcea for further<br />

reccxzmendationo or acceptance.<br />

b. A draft copy of the WA Bylaw? was discussed in detail.<br />

The recanmended changer will be incorporated in the redraft of the<br />

Bylaws which wtll be submitted to the Servicee. It fe anticipated<br />

that a final copy will be presented to the membership at the 1965<br />

MTA conference.<br />

c. Major Frank I.. McLanethan of the Air Fcrce was elected<br />

to the Chairmanehip of the 1965 MTA Steering Colmnfttee.<br />

181<br />

..-----


-.<br />

Item Writers Aptitude Test Development Coornittec <strong>Report</strong><br />

Date of Meeting: 21 October 196t.<br />

Place of Meeting: EfCA Conference, Fort Benjamin HarrLdon, Indianapolis,<br />

Indiana.<br />

Committee Members: Mr. William W. !kr!ce, USA, Chainon<br />

Lt Harvey C. Gregoire, USAF<br />

Mr. fn:,rlte A, Nudoon, US?4<br />

I:r. John D. Kraf t, USA<br />

Hajor Joe R. Shafer, USAF<br />

#r. William M. Minter, USA<br />

hmary of Discussion:<br />

a. All services rsprejented gave a brief description of hov their<br />

test item writers are selected, None of the service8 currently u8es er.y<br />

special selection procedures other than subject-mhttar knowledge. The<br />

coernittee members felt that Subject-matter background was not a serious<br />

problem since the itcm uritars befng designated for test item construction<br />

duties are generally adequate in this respect.<br />

b. In general) the item urttero in all services perform skmilar<br />

duties. The Navy and Air Porte have their ftem writera come to a central<br />

location here they work directly with tC8t speclalist8. The Amy operates<br />

on a more decentralized basis in that the item writers are located mainly at<br />

Army Sertlice Schools and construct test itomr bared on wrtttcu speclffcotionr<br />

sent to them by the US Army Enlisted Evaluatior, Center. Test Sp.?Ciatt~tS<br />

from the latter locetfon make TDY trips to the item-writing agencies to<br />

coordinate the test development efforts,<br />

c. The ccmmittes felt that the followfng two major areas should be<br />

inveetigated as being most likely to provide valid fact ~8 which would<br />

asslat in screening the better from the poorer test item writers:<br />

(1) Personal hfatory items.<br />

(2) Test results obtained from measuring factoru, such a8 English<br />

usage, intelligence, and analytical reading ability. ,<br />

d. Some “brain storming” vas done by the cora3fttee to get some fndfcation<br />

of the types of personal history items or teets that might be used<br />

to asse8s item writera’ aptitude. However, the committee felt that a<br />

thorough and ccmprehenstve analysfe would need to be made of the f tern writer<br />

duties before the best groupfng of potential personal history item8 and<br />

tests could be compiled for use in fretting up an experimental desfgn appropriate<br />

for conducting a validation study.<br />

1. -<br />

.<br />

182<br />

_ . _ _.<br />

_.._.-_--.. -_--- -.--.--<br />

- - _^<br />

.-<br />

. .<br />

-<br />

.


e. Lt Gregolre, US&?, indicated that the S*KI (Specialty Knowledge<br />

Test) Branch at the Personnel Research Laboratory at Lsckland AI% was<br />

very much interested in the asseamnent of ?tes writer aptitudes. RC<br />

also fndicatcd that sufficient data and processing support very likely<br />

would be available at Lacklnnd AFB to conduct a preliminary study related<br />

to assessment of itm writers,<br />

f. Tentative plans wer


.<br />

i .<br />

--<br />

t<br />

!<br />

Ens Harold AdaL-an<br />

Hr. fied AlRrad<br />

Dr. Neal B. A~zdragg<br />

Lt Thumars 11. Atchley<br />

Mr. 841~ II. Baker<br />

Hr. Rqmond V. A&at<br />

Mr. Vernon H. Begge<br />

pi-r. Walter W. Rirdaill<br />

Dr. Warren S. Bluxmfald<br />

Mr. M&eel J. Bodi.<br />

Lt Barbara Bole<br />

Hr. John S. Brand<br />

Wr. Claude P. Brfdgee<br />

Mr. Donald A. Brown<br />

2d Lt Martin S. Brown<br />

Hre. Habal 0. Brunner<br />

ROSTER OF CONFEREEES<br />

Yx. Franklin S. Buckwmlter US Amy Quertermaatcr School<br />

Xaj Robert A. Burgesa<br />

Mr. Anthony Cmciglia<br />

u??lT ATTACHED<br />

- __yI--<br />

‘tlS Coset msrd Homdquartera<br />

US Coast &ard Training Cmto?:<br />

US Amy Security Agency<br />

us Am] w1itGy Polfce corpm<br />

US Naval ExDmFninS Cente:<br />

US Army Enlisted Evaluation Center<br />

6570th Personnel Research bborutowry<br />

US Air Force<br />

US Haval Exminlng Center<br />

US Naval Em~ining Center<br />

US Naval Peraonncl Research Activity<br />

Sm Diego<br />

US Amy Enli.ett:l Evaluation Center<br />

US Naval Exmlning Center<br />

US Army Enlisted Evaluation Center<br />

US Amy Enlisted Evsluetton Center<br />

US Amy Mr.dical Service, Fitzeinmons<br />

US Amy Enlf.sted Evaluation Center<br />

Headquarters, UZ Air Force<br />

US Army Enlfrrted Evaluation Center<br />

US Army Security Agency<br />

US Amy Enlisted Evaluation Center


_’<br />

. ‘-<br />

-.<br />

I<br />

I<br />

:<br />

e .<br />

I. .I<br />

--.. .._ _<br />

L<br />

C<br />

.<br />

c<br />

--_I.---<br />

. ..-A.<br />

Capt. JCE?C?B B. Carpenter<br />

Mr. Charles E. Caosidy<br />

:tt co1 c. f. chafenon<br />

Mr. Tbamz E. Chandler<br />

!XC William Q. Chcnn<br />

ILt Co1 Kent J. Ccllfngr<br />

lb. John W. Credlford<br />

Maj David P. Culclasure<br />

Cdr R. J. Dahlby<br />

M~aj Donald L. Dimand<br />

l.at Lt Duncan L. Dieterly<br />

Co1 Jamer C . Donaghey<br />

Dr. Henry J. Duel<br />

Mr. Erling A. Dukrrrchein<br />

?CSgt Raymond R. Durand<br />

1.6t Lt Jane* H. Durden<br />

Mr. Bernard J. Foley<br />

Mr. Step‘.:+- W. Pctlr<br />

Mr. John L. Pinucrnr<br />

Lt C. E. Csngenbneh<br />

Mr. Ronald K. Goodnight<br />

Cdt Robert J. Gray<br />

____.- _.... -.<br />

185<br />

6570th Personnel Research Laboratory<br />

US Air Force 0<br />

US Army Enlitited Evaluation Center<br />

US Army Enlisted Evaluation Center<br />

US hrmy Southeastern Signal School<br />

US Army Cmbat Survefllance Schocl<br />

US Army Enlisted Evaluation Center<br />

US Naval Zxamining Center<br />

US Amy Hedical Field Service<br />

US Coast Guard Training Center<br />

US t??rfne Corpr Inotftuta<br />

6570th Personnel Raecarch &sborrtory<br />

US Air Force<br />

US Army Enlfetcd Evaluation Center<br />

Headquarters, US Air Force<br />

US Naval Exmining Canter<br />

US Amy Security Agency Training Center<br />

US Army Ordnance Guided Hierilc School<br />

US Amy Security Agency Trafnlng Center<br />

6570th Personnel Research Laboratory<br />

US Air Force<br />

US Army Enlfrted Evaluation Center<br />

US Army Enliertb :;:lurtion Center<br />

US Naval Examining Center<br />

. -, -..-.. _ _<br />

__, y


1<br />

/<br />

. . .<br />

I<br />

.<br />

Cape Harry H. Craer, Jr.<br />

1st Lt Harvey Cregorie<br />

Ltjg Clyde A. Gronewold<br />

1st Lt Thomar H. Guback<br />

Mr. Robert L. Cup<br />

(Cdr Frederick J. Hancox<br />

Mr. Clayton B. Haradon<br />

I,t ‘Co1 Roy E. HaIrin, Jr.<br />

Capt R. H. Hayee<br />

F!r. Jock E . Hohrelter<br />

tfx. Fred 8. Hona<br />

Mr. John C. Houtr<br />

Hr. John J. Hcjbell<br />

Hr. Charles A. Hudson<br />

Hr. Clifford 2. Hutrlcy<br />

Mr. William L. Jackson<br />

HI :. E. C. Johnron<br />

Hr. L. W. Johnston<br />

Lt jg Katherine Kadenacy<br />

T.Edr Lawrence R. Kilty<br />

2d Lt Lloyd 0. Kimery<br />

Mr. Albert Kind<br />

UFXT ATTACREZD<br />

us Navy (Roti;ed)<br />

6570th Personnel Research Laboratory<br />

US Air Force<br />

US Navel Examining Center<br />

US Amy Defense Xnformotion School<br />

US Naval Examining Center<br />

US Coast Guard Headquarter6<br />

6570th Pcreonnel Research Laboratory<br />

US Air Force<br />

186<br />

Randolph Air Force Base<br />

US Naval Examining Center<br />

US Army Ealieted Evaluation Center<br />

US Army Enlisted F,valu:tFon Center<br />

US Naval Exemfning Center<br />

US Army AzzYor School<br />

US Naval Exeminlng Center<br />

US Army Signal School<br />

US Army Aviation School<br />

US Army Enlisted Evaluation Ctnter<br />

US Naval Examlnlng Center<br />

US Naval Examining Center<br />

US Naval Security Group Headquarter6<br />

US A- Training Ctnter, Englnoars<br />

US Army Combat Surveillance School<br />

. .<br />

. .<br />

%<br />

.<br />

_


_. . . i<br />

I<br />

.<br />

c<br />

Lt Co1 Albert S. Knauf<br />

Xr. Richard S. Kneirel<br />

Xr. John Kraft<br />

Cdr Ger?e L. Lane<br />

Mr. Larry J. LeBlanc<br />

LCdr Jack D. Lee<br />

Dr. C. L. John Lagere<br />

1st Lt Robert H. Lennevflle<br />

Lt Alexander A. Longo<br />

Xr, Charles J. Macaluao<br />

Capt Jack H. Harden<br />

Lt David R. Hnrkep<br />

Capt Joreph P. Kartin<br />

Hr. Curtis D. KcBrlde<br />

Hr. NflliuJ X. Hinter<br />

Dr. Joseph E. Harsh<br />

Mr. Ieadora J. Nevolpn<br />

Hr. W. Alcn Nicewander<br />

Ltfg Richard L. Olsen<br />

rtrj Herr11 R . Ower,<br />

1st Lt Arnold J. Pals<br />

ICdr Ralph Palverky<br />

-___ ..-<br />

. .<br />

6570th Personnel. Research Laboratory<br />

US Air Force<br />

US Amy Chamicnl Center and School<br />

OS Amy Enlfsted Eveluatlon Center<br />

Bureau of Naval Peroonnel<br />

Lackland Air Force Base<br />

US HAVA~ Examfning Center<br />

US Army Security Agency Training Center<br />

Navy School of Horic, Army Element<br />

US Naval Air <strong>Technical</strong> Training Center<br />

US NAVR~ Examining Center<br />

US Army Judge Advocate General School<br />

US Coast Guard Training Cer.ter<br />

US Coaat Guard Training Center<br />

US Axuy Artillery and Xiaeile School<br />

US Array Chemical Center and School<br />

6570th Personnel Research Laboratory<br />

US Air Force<br />

6570th Personnel Research LPboratory<br />

US Air Force<br />

US Army Enlirted Evaluation Center<br />

US Coart Guard Training Center<br />

US Amy Bnlieted Evaluation Center<br />

US Army Medical Center, Walter Reed<br />

Naval Examination Center, Royal Canadian<br />

N&V<br />

107<br />

.<br />

.<br />

_ I .- . . -<br />

I<br />

_.-- I .-__I .*.-a<br />

._- --- _ . . ^ e-e... .--<br />

1


.<br />

‘.<br />

. &<br />

.<br />

e<br />

.<br />

Fir. John J. Parke<br />

Mr. J. E. Ptrtingtcm<br />

Capt Kennuth A. Petri&<br />

M&j Jams N. Payne<br />

Hr. Henry W. Pepin<br />

Lt David L. Popple<br />

Mr. William Ii. pitman<br />

Uaf Joseph T, Polanrki<br />

Capt Carl R, Powers<br />

Mr. Frank R. Prfcs, Jr.<br />

Lt Co1 Robert A. Rerrnnyder<br />

\ Msj Clinton D. Regelia<br />

.<br />

I<br />

-..<br />

:<br />

-- ..---<br />

Mr. John H. Roth8<br />

Hr. Jack Rubak<br />

Hr. Carl Rudlnekl<br />

let L t Jumee L. Rueaell<br />

Maj William H. Sallejr<br />

Cept Clsrance D, Sspp<br />

Mrs. Genevieve K. Schulter<br />

Hoj Joe R. Shaafer<br />

Mr. Jean B. Sheppard<br />

Hr. Edwin C. Shtrkey<br />

SSgt Predrick J . Shunk<br />

.<br />

,<br />

,<br />

UN IT AYL’AC MD<br />

US Amj Ordnmce Guided Memile School<br />

US Army Enll$ted Evaluation Center<br />

US Amy Combat SurveilLaxe School<br />

US At-my Defenlre Atomic Support Agency<br />

US Amy Defame AtmEc Support Agency<br />

US Coast Guard Reserve Trainfng Center<br />

US Army Ordnance Center and School<br />

US Army <strong>Military</strong> Police School<br />

US Defense Atomic Support ABWICY<br />

US Army Enlisted Evaluation Center<br />

US Army Enlirttd Evaluation Center<br />

US Amy Defenm Information School<br />

Dept of Wavy, Bureau of Naval Personnel<br />

US Asq Dafanes Inforjletion School<br />

US Ar~rg S’qnal Cen.tar and School<br />

US Amy Infantry School<br />

US Army Intelligence School<br />

US Amy Air DL’fenoe School<br />

US laval Examining Center<br />

6570th Perronnel Rescsrch Laboratory<br />

US- Air Force<br />

188<br />

US Army Southeastern Signal School<br />

US Amy Enliattd Evaluation Center<br />

US Army <strong>Military</strong> Police School<br />

I<br />

‘.<br />

.<br />

. -<br />

,<br />

Y<br />

__. . ---<br />

.<br />

4<br />

.


.___....--<br />

i<br />

Mr. Willi6m P. SFm<br />

Capt Loren K. Smith<br />

j<br />

. Mr. Jane6 W. Smith<br />

1<br />

!<br />

) *<br />

-<br />

Mr. MAxon Ii. Smith<br />

Xdr Doncnld H. Tart<br />

Dr. June6 D. Teller<br />

ICdr Prrncea S. Turner<br />

Hr. Vern W, Urry<br />

Dr. Raymond 0. Weldkoetter<br />

Mt. FrUICi6 B . Waleh<br />

Mr. William W. W6nce<br />

Dr. Donald L. W666<br />

Mr. kdalbdrt U. Wafabrod<br />

Lt Berl R. Wll.li6ms<br />

Ens Thomas H. Wilson<br />

c Hr. C66tier S. Wlniwi66<br />

Dr. Michael A. Zaccarirr<br />

_ ._ ._ -, - _ . -<br />

.<br />

_.<br />

. .<br />

189<br />

UNIT ATTACHED<br />

US Arq Air Dafennse School<br />

US Army Intelligence School<br />

US Amy Ordnance Guided Miasfla School<br />

US 8rzz-y Ordnwce Gufdod Hiceile School<br />

Royal CanadiAn N6vy<br />

Headquartera, US Air Force<br />

US Coa6t Guard Training Center<br />

US Army Enlirted Evaluation Center<br />

US Army Enlieted Evuluation Center<br />

US Army Security Agency Training Center<br />

US Army Enlisted Evaluation Centur<br />

US Arq EnAl. l’6t6d Evaluatfon Center<br />

US Army Ccrpa of Engineer6<br />

US Coast Guard Training Center<br />

US Coast Guard Training Center<br />

US Naval ExaininLng Center<br />

Lockland Air Porca Baas Training Center<br />

.


6..<br />

,<br />

. .<br />

1%<br />

+ ,<br />

,*<br />

___<br />

.<br />

‘. *<br />

, ~ I<br />

, -I . . ;<br />

/ . . ‘x 4. . - . 1; \ .‘.), *.


UNCLASSIFIED/UNLIMITED<br />

PLEASE DO NOT RETURN<br />

THIS DOCUMENT TO DTIC<br />

EACH ACTIVITY IS RESPONSIBLE FOR DESTRUCTION OF THIS<br />

DOCUMENT ACCORDING TO APPLICABLE REGULATIONS.<br />

UNCLASSIFIED/UNLIMITED<br />

_. - .- ._. .-

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!