Evaluating Country Programmes - OECD Online Bookshop
Evaluating Country Programmes - OECD Online Bookshop
Evaluating Country Programmes - OECD Online Bookshop
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
<strong>OECD</strong> 1999<br />
<strong>Country</strong> Programme Evaluation: A State of the Art Review<br />
It is necessary to examine each agency’s own definition of the terms as there<br />
is in most cases room for debate about whether a given item should be<br />
regarded as (for example) an immediate objective or an overarching goal.<br />
The Netherlands, for example, does not mention impact: however, in defining<br />
effectiveness as “contribution to [the country’s] macro-economic self-reliance,<br />
poverty alleviation, and the sustainability of project results”<br />
(Netherlands/India, 1994: 50) it seems that it is talking not only about effectiveness<br />
but also impact (and, in this case, sustainability too).<br />
ii) There are problems in using logframe analysis, designed primarily as a tool<br />
for project planning, to evaluate a complex country programme. A country<br />
programme includes numerous, often imperfectly articulated or even contradictory<br />
goals, to be achieved through a range of instruments, sectors and<br />
channels. Few donors have attempted to apply criteria of efficiency, effectiveness<br />
or impact to programme aid (White, 1996); technical assistance is<br />
also hard to evaluate in logframe terms. This is a problem when attempting<br />
to assess the relative contribution of programme aid and project aid components<br />
of a given country programme (the problem of comparing “apples and<br />
pears”: Maxwell, 1996), and when a country programme is predominantly<br />
implemented as programme aid.<br />
Thus, while there may be value in drawing on logframe-based concepts of the<br />
relationship between criteria in country programme evaluation, there is also a need<br />
to define all criteria in an explicit, logical and consistent way that is amenable to<br />
measurement (or estimation). The intrinsic “evaluability” of activities may vary considerably<br />
(IDB, 1998: 11-12).<br />
In an effort to introduce consistency, DAC has provided a list of standardised<br />
definitions of evaluation criteria (Box 2.3).<br />
On the basis of these standardised definitions, we found the following criteria<br />
presented in the descriptions of methodology in the CPEs we reviewed. It should<br />
be noted, however, that in many cases a criterion was mentioned and defined but<br />
not subsequently used to any significant extent in the analysis of the country programme<br />
which followed.<br />
Some studies also used other criteria based upon the particular circumstances<br />
of the partner country. In evaluating aid given via multilaterals to Cambodia, a SIDA<br />
study (not a CPE) addressed “additionality”: as many donors initiated aid to Cambodia,<br />
SIDA wished to ensure that it gave aid to projects that would not have found<br />
another donor (SIDA/Cambodia, 1995). This could arguably be seen to be a subset<br />
of the criterion of “relevance”.<br />
From project to country programme<br />
<strong>Country</strong> programme evaluations vary in the degree to which they attempt to<br />
provide an evaluation of the country programme which is “more than the sum of the<br />
71