aceUVi
aceUVi
aceUVi
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
espondents are conscious; 2) there is limited comparability across events and<br />
locations (ie, the programmes being assessed are different, as are the participants<br />
and community contexts); 3) surveys immediately following events fail to capture<br />
effects that unfold over time.<br />
The primary focus of the literature on post-event survey research has been developing<br />
tools to assist arts and cultural organisations in gathering high quality<br />
audience feedback for internal accountability purposes. The larger question of the<br />
role of post-event survey data in assessing public sector policy is not addressed<br />
in the literature, except in the work of Chappell and Knell (2012) and Bunting<br />
and Knell (2014). If anything, the literature raises questions as to the plausibility<br />
of aggregating survey data across organisations and artforms, due to the highly<br />
personal and situational nature of impact, and because of differences across the<br />
forms themselves (Belfiore and Bennett 2007). Additional research and debate is<br />
required to better understand how and when it is meaningful to aggregate self-reported<br />
participant impact data across organisations and artforms.<br />
Qualitative post-event research<br />
Qualitative methods have the advantage that they allow informants to focus their<br />
reflections on the areas that are most significant to them. Rather than defining<br />
constructs in advance, researchers can allow interviewees to express themselves<br />
in their own terms and subsequently derive the most relevant categories of<br />
responses inductively (Radbourne 2009, 2010a, Walmsley 2013). Thus, Foreman-<br />
Wernet and Dervin (2013) found that their informants discussed negative or<br />
mixed responses, which might not have come to light in quantitative surveys. As<br />
with formal surveying, however, qualitative methods of inquiry can only capture<br />
aspects of an experience of which respondents are aware and that they are able to<br />
articulate.<br />
Whereas survey methods can be expected to yield consistent results over time<br />
and thus generate a stable basis of knowledge, the results of qualitative studies<br />
are not replicable, independently verifiable, or refutable. Since the conclusions<br />
drawn by one qualitative study are not disproven by contradictory findings in<br />
another, there is no inherent mechanism that weeds out research that is of substandard<br />
quality. One must therefore assess the integrity of each study by paying<br />
close attention to the research design and methodology, which can be challenging<br />
for lay readers. Nonetheless, several researchers have stressed the important role<br />
qualitative studies play in contextualising numerical data (O’Brien 2010, Klamer<br />
2004, Holden 2006). Narrative accounts can tell us why people value cultural experiences<br />
and what those experiences mean to them, rather than just measuring<br />
to what extent they were affected. With the advent of more sophisticated textual<br />
Executive Summary 13<br />
UNDERSTANDING the value and impacts of cultural experiences