Evaluating Country Programmes - OECD Online Bookshop
Evaluating Country Programmes - OECD Online Bookshop
Evaluating Country Programmes - OECD Online Bookshop
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
<strong>OECD</strong> 1999<br />
<strong>Country</strong> Programme Evaluation: A State of the Art Review<br />
Box 2.4. Examples of different categorisation systems<br />
used to judge country programme performance in the UNDP<br />
Different UNDP evaluations have used different systems.<br />
Sri Lanka: self-evaluation of overall programme performance using scoring<br />
In evaluating components of the Sri Lanka programme, the evaluators asked<br />
country programme staff to score UNDP performance in 13 "sub-themes" of the programme<br />
(in terms of contribution to, for example, "high and sustained economic<br />
growth" or “a leaner, more efficient and service-oriented public administration")<br />
according to the following guideline:<br />
5. Made substantial progress.<br />
4. Made some progress.<br />
3. No difference made.<br />
2. Negatively affected progress somewhat.<br />
1. Negatively affected progress substantially.<br />
The scores given were then averaged for each sub-theme (UNDP/Sri Lanka,<br />
1994: 17, 46-7).<br />
Myanmar: external evaluation using classification<br />
In an earlier evaluation of the programme in Myanmar, 51 projects were each<br />
classified as satisfactory, unsatisfactory or partially satisfactory/partially unsatisfactory<br />
(UNDP/Myanmar, 1992: 29). In this case, the judgement was made by the evaluation<br />
team rather than the implementing staff. Broad conclusions about country<br />
programme effectiveness were guided by the proportion of projects which were<br />
judged (for example) successful, unsuccessful or partially successful.<br />
If projects are classified or scored, the key question is who provides the classification<br />
or scoring. An individual evaluator or small team of evaluators making their<br />
own decisions about project performance (effectiveness or impact) has the advantage<br />
of greater consistency (i.e. their subjective judgements are at least consistent,<br />
so that “good” means the same in all cases). However, evaluators will generally not<br />
be able to acquire the depth of knowledge required to make an informed<br />
judgement of each and every element within the country programme portfolio, at<br />
least not within the timeframe of the typical CPE. More commonly, they ask those<br />
involved in the country programme to classify or score those projects with which<br />
they are familiar.<br />
79