Radical innovation: crossing knowledge boundaries with ...
Radical innovation: crossing knowledge boundaries with ...
Radical innovation: crossing knowledge boundaries with ...
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Metrication would not be necessary as a component of interdisciplinary <strong>innovation</strong>, if<br />
it were not for the need for sponsors to carry out formative evaluation to prioritise<br />
funding investments and identify the value of likely outcomes. Once a funding<br />
decision has been made, there is no further value from metrication, unless it is to<br />
provide ways of characterising unanticipated outcomes as they arise.<br />
There have been some attempts to assess the potential impact of increased research<br />
metrication on interdisciplinary research, where the assessment evaluates metrics<br />
themselves for consistency across disciplines and into interdisciplinary work (Levitt<br />
& Thelwall 2008, Adams, Jackson & Marshall 2007). However, we believe that the<br />
terms of reference for such studies have been too narrow to date, and do not take into<br />
account most of the dynamics that appear to be important in our own findings. For<br />
example, although Levitt and Thelwall found that multidisciplinary publications in<br />
science and technology were only half as likely to be cited as monodisciplinary<br />
publications, Adams et.al. defined interdisciplinarity so narrowly that major<br />
challenges to organizational and <strong>knowledge</strong> <strong>boundaries</strong> do not arise (for example, by<br />
only considering science and technology disciplines, or by treating statistics and<br />
applied statistics as different disciplines across which citations would be regarded as<br />
evidence of interdisciplinarity).<br />
Nevertheless there is a clear connection between these public value measures, and<br />
evaluation of new recruits to a discipline, via school and degree examinations. Almost<br />
all researchers are selected for professional research careers because they have been<br />
‘metricated’ via earlier examinations, and these have been constructed in accordance<br />
<strong>with</strong> the <strong>knowledge</strong> of particular disciplines.<br />
It is interesting to speculate whether quantitative social network analysis of the kind<br />
developed by Burt could be used as a valuable form of metrication for the evaluation<br />
of capacity building. However Marilyn Strathern warns against this confusion of<br />
means and end. Metrication is an essential component of the construction of<br />
disciplinary elites. As inter-disciplines become disciplines, suitable metrics for that<br />
discipline are likely to evolve.<br />
Tom Inns recommends developing a portfolio of metrics, to account for the necessary<br />
combination of short-term and long-term outcomes. The elements of this portfolio<br />
might reasonably take into account all of the dynamics described in section 5 on<br />
managing interdisciplinary teams, and in particular section 5.8 on evaluation.<br />
However, such ‘metrics’ would be very unlike those we work <strong>with</strong> at present. It is<br />
important to recognize how many participants in this research felt that current policy<br />
regarding metrication of research was fundamentally anti-innovative.<br />
Case study: Interdisciplinarity in Medicine<br />
There are some situations in which a characteristic set of obstacles appear - we<br />
illustrate this using specific examples from medical practice. There are are, of course,<br />
other fields <strong>with</strong>in medicine that do exhibit the kinds of <strong>innovation</strong> that we have<br />
described elsewhere in this report.<br />
Innovation and Interdisciplinarity 77