09.12.2012 Views

I__. - International Military Testing Association

I__. - International Military Testing Association

I__. - International Military Testing Association

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Implementation of Content Validity Ratings<br />

in Air Force Promotion Test Construction<br />

Carlene M. Perry<br />

United States Air Force Academy<br />

John E. Williams<br />

Paul P. Stanley II<br />

USAF Occupational Measurement Squadron<br />

The USAFOMS has implemented a procedure in which subject-matter experts<br />

(SMEs) rate the content validity of individual test questions on Air Force<br />

promotion tests. This paper describes that procedure and assesses its impact<br />

upon test content and the perceptions of those involved in test,developm.ent.<br />

Specialty Knowledge Tests (SKTs) are loo-item multiple-choice tests which<br />

measure airmen's knowledge in their assigned Air Force specialties. Promotion<br />

to the ranks of staff sergeant (E-5) through master sergeant (E-7) is<br />

determined solely by relative ranking under the Weighted Airman Promotion<br />

System (WAPS). Each airman receives a single WAPS score, the sum of six com-<br />

ponent measures, with the SKT accounting for UP to 22% of the total. Since<br />

the other components generally yield little variance among individuals, the<br />

SKT is often the deciding factor in determining who gets promoted.<br />

SKTs are written by teams of senior NCOs with the assistance of USAFOMS psychologists.<br />

They are constructed using the content validity strategy of validation.<br />

The following components ensure a direct and close relationship<br />

between test items and important facets of the specialty being tested: 1)<br />

the specialty training standard, an Air Force document which identifies the<br />

primary duties and responsibilities in a specialty, 2) CODAP-based occupational<br />

analysis data collected by USAFOMS, and 3) the expertise of the SMEs<br />

themselves.<br />

Content validity is thoroughly documented for the more than 400 SKTs in the<br />

USAF inventory. However, the documentation is predominantly qualitative<br />

rather than quantitative in nature, as is the norm with tests based on this<br />

strategy. Test developers at USAFOMS felt that a quantitative means of assessing<br />

content validity would be useful to improve their feedback to test<br />

writers and to make it possible to study test resulfs longitudinally to help<br />

identify problem tests.<br />

Lawshe (19751 developed just such an approach, the first to focus on content<br />

validity in a quantitative, rather than a qualitative manner. His method<br />

called for a panel of subject-matter experts (SMEs) to independently rate<br />

test items using the following scale:<br />

Is the skill (or knowledge) measured by this test question:<br />

Essential (21,<br />

Useful but not essential Cl>, or<br />

- Not necessary (01,<br />

for successful performance on the job?<br />

He then used a test of statistical significance with the content validation<br />

panel’s ratings as a basis for eliminating items from a test item pool.<br />

Lawshe’s procedure has been applied by a number of investigators in a variety<br />

of situations. Distefano, Pryer, and Craig (1980) used his content valida-<br />

235<br />

.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!