09.12.2012 Views

2003 IMTA Proceedings - International Military Testing Association

2003 IMTA Proceedings - International Military Testing Association

2003 IMTA Proceedings - International Military Testing Association

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

276<br />

we surveyed all of the employees. Whenever the SME panels created a new work role or merged<br />

existing work roles, they were required to identify a suitable sample of employees to complete<br />

the survey. Each survey gathered respondent background information and ratings of the core<br />

competencies. For each competency, a dichotomous (yes/no) format item asked whether the<br />

competency is used. When the respondent indicated “yes,” the survey asked three additional<br />

questions: the competency’s importance to performing the work, the extent to which it is needed<br />

upon hire, and the amount of time that the respondent spends using it. The importance, hire, and<br />

time items used 5-point Likert scales, with anchors ranging from “Strongly Disagree” (1) to<br />

“Strongly Agree” (5). Because the survey was electronically-deployed, validation rules<br />

prevented out-of-range values.<br />

Next, we weighted the response data at two steps. The first weighting was made at the<br />

work role level. The purpose of this first weighting was to adjust the PAB ratings to contribute<br />

50 percent of the final weighted response. For each work role, we identified the overall<br />

importance for each competency by calculating the weighted mean and standard deviation of<br />

ratings for need at hire, frequency, and importance across all the incumbent raters for that work<br />

role and its PAB (with the PAB ratings weighted as previously discussed). After calculating the<br />

weighted overall importance ratings, we sorted the competencies within competency type<br />

(knowledge, skill, cross-occupational skill, and tool) in decreasing order of importance (mean)<br />

and increasing consensus about importance (standard deviation). The second weighting, which<br />

was made at the occupation level, was designed to provide greater weight to competencies that<br />

are identified in a large number of roles. Therefore, we weighted each competency by the<br />

number of work roles in which it appeared. For example, if a competency appeared in 5 work<br />

roles, it was weighted by 5; if a competency appeared in only 1 work role, it was not weighted.<br />

This weighting procedure was done to ensure that the most “critical” competencies were in fact<br />

used by a large percentage of the NIMA workforce.<br />

We then used Chronbach’s alpha to calculate the overall degree of inter-rater reliability<br />

(or consensus) within each work role. Cronbach’s alpha represents the degree to which raters<br />

provided consistent evaluations of the importance of each competency. Cronbach’s alpha is<br />

based on both the number of raters, as well as the similarity of the ratings included in calculating<br />

it. Consequently, low values may also be due either to small sample size alone, low consensus<br />

alone, or both low consensus and small sample size. Values for Cronbach’s alpha range from 0.0<br />

(indicating no consistency) to 1.0 (indicating perfect agreement). Values under 0.50 may suggest<br />

that the raters disagreed about the importance of the competencies they rated, and possibly that<br />

the work role may require additional review. However, low values can also result when the work<br />

role had only a small number of reviewers. Conversely, high values indicate agreement among<br />

the reviewers about the importance of the competencies they rated, suggesting that the work role<br />

more accurately describes the work.<br />

Finally, we used the Jaccard similarity coefficient to indicate the degree to which pairs of<br />

work roles had the same competencies identified as being essential for performing work they<br />

describe. Specifically, we considered values for the Jaccard coefficient greater than 0.80 as<br />

suggesting that further review should be considered. As before, statistical results were combined<br />

with expert judgments to determine whether or not individual pairs of work roles should be<br />

combined. Before the results data could be integrated into PeopleSoft, they needed to be<br />

45 th Annual Conference of the <strong>International</strong> <strong>Military</strong> <strong>Testing</strong> <strong>Association</strong><br />

Pensacola, Florida, 3-6 November <strong>2003</strong>

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!