08.01.2015 Views

Learning from Nine Examples of Peacebuilding Evaluation

Learning from Nine Examples of Peacebuilding Evaluation

Learning from Nine Examples of Peacebuilding Evaluation

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Key findings <strong>of</strong> the case studies were: practitioners highly valued the training<br />

and ongoing specialist support; parents welcomed the opportunity for their<br />

children to learn about respecting difference; committed leadership within<br />

the setting where the curriculum was delivered was crucial; and integration<br />

<strong>of</strong> the various aspects <strong>of</strong> the project across different settings was important<br />

to successful implementation.<br />

Considerations on the Methodology<br />

Strengths<br />

• While the Evidence Summit was not designed to produce a rigorous peer<br />

review <strong>of</strong> the evaluation methodology, the RCT discussed here appears to<br />

have been carefully constructed and implemented. The methodology included<br />

true randomization, sufficient power, and credible instruments to measure<br />

the change in attitudes among the children.<br />

• The benefit <strong>of</strong> this methodology, according to Early Years, was that it produced<br />

a simple, powerful measure <strong>of</strong> impact, which was important for their<br />

engagement with donors and with skeptics who believed that introducing<br />

these issues to young children would be counterproductive.<br />

• The inclusion <strong>of</strong> the qualitative research broadened the appeal <strong>of</strong> the<br />

research and generated crucial buy-in <strong>from</strong> staff members, parents, and educators.<br />

The results <strong>of</strong> the qualitative research, Early Years reports, were more<br />

important for stakeholders on a day-to-day basis.<br />

25<br />

Challenges and Pitfalls<br />

• The first key challenge was the financial and other costs <strong>of</strong> the RCT. Due<br />

to the disruption caused within the schools and the complexity and cost <strong>of</strong><br />

the research, the evaluation was necessarily a one-<strong>of</strong>f initiative. It was not a<br />

process that could feasibly be repeated at regular intervals to create ongoing<br />

learning. The organization continues to evaluate their programs but now<br />

primarily uses more qualitative, participatory approaches.<br />

• The use <strong>of</strong> external researchers created problems. Early Years had to<br />

balance the demands <strong>of</strong> the research with their own community development<br />

principles and the dynamic nature <strong>of</strong> the projects being assessed. In<br />

addition, the researchers needed to do their own learning <strong>of</strong> the work <strong>of</strong> Early<br />

Years. Finally, Early Years noted that the design limited research capacity<br />

building within the organization.<br />

• Early Years reported that the qualitative research did not capture the richness<br />

<strong>of</strong> the program, nor the full impact, especially with parents. The bulk <strong>of</strong> the<br />

resources for the initiative went to the RCT. The qualitative research was not<br />

given equal emphasis, which may have created some <strong>of</strong> these shortcomings.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!