02.12.2012 Views

Intellectual Capital Reporting for Universities - ARC systems research

Intellectual Capital Reporting for Universities - ARC systems research

Intellectual Capital Reporting for Universities - ARC systems research

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Intellectual</strong> <strong>Capital</strong> <strong>Reporting</strong> <strong>for</strong> <strong>Universities</strong>: Conceptual background<br />

and application within the reorganisation of Austrian universities<br />

Dr. Karl-Heinz Leitner<br />

Department of Technology Policy<br />

Systems Research Technology-Economy-Environment<br />

<strong>ARC</strong> Seibersdorf <strong>research</strong>GmbH<br />

A-2444 Seibersdorf<br />

Paper prepared <strong>for</strong> the Conference<br />

”The Transparent Enterprise. The Value of Intangibles.”<br />

Autonomous University of Madrid<br />

Ministry of Economy<br />

November 25-26, 2002, Madrid, Spain<br />

Abstract: European universities are faced with numerous challenges caused by the<br />

political aim to harmonise the different national university <strong>systems</strong> and new <strong>research</strong><br />

modes and by the claims of various stakeholders such as the industry and society. In many<br />

countries universities are increasingly provided with greater autonomy regarding their<br />

organisation, management, and budget allocation, which requires new management and<br />

reporting <strong>systems</strong>. As organisations that are mainly funded by the public sector, they are<br />

also faced with an increased demand <strong>for</strong> transparency and accountability. <strong>Universities</strong> are<br />

knowledge producers per se, their most important output is knowledge incorporated in new<br />

<strong>research</strong> results, publications and educated students. A university’s most valuable<br />

resources are the <strong>research</strong>ers and students with their relations and organisational routines.<br />

These resources can be interpreted as intangible assets, even though the term has so far not<br />

been used within the context of universities. The instrument of intellectual capital reporting<br />

and methods <strong>for</strong> valuing intangibles seem to be a rich foundation <strong>for</strong> the development of<br />

the necessary new management and reporting <strong>systems</strong>. The Austrian Ministry of Education<br />

has decided to study the potential of intellectual capital reporting and will introduce this<br />

new instrument in the course of the reorganisation of the Austrian universities in the future.<br />

The conceptual framework will be presented in this paper, embedded in the international<br />

discussion of new management and valuation <strong>systems</strong> <strong>for</strong> universities.<br />

Key words: Research valuation, universities, evaluation, education, <strong>Intellectual</strong> <strong>Capital</strong><br />

Report, <strong>research</strong> management, science system, per<strong>for</strong>mance management, new public<br />

management<br />

1


1. Introduction<br />

In most European countries universities are continuously faced with various new<br />

challenges caused by political initiatives, an altered economy and society as well as new<br />

<strong>research</strong> modes. A major challenge <strong>for</strong> universities, driven by European-wide political<br />

activities, such as the Bologna declaration and its mission to create the European Higher<br />

Education Area, is the harmonisation of the individual national university <strong>systems</strong>. The<br />

aim of this declaration is to achieve greater compatibility and comparability of the higher<br />

education sector. Measures such as the establishment of a system of credits as a means to<br />

promote student mobility, the adoption of a system of comparable degrees as well as the<br />

promotion of quality assurance are among the first initiatives (Salamanca 2001).<br />

In recent years in many countries universities have been provided with greater autonomy<br />

regarding their organisation, management, and budget allocation. This has also raised also<br />

questions regarding their basic functions and tasks within society, which are, <strong>for</strong> instance,<br />

expressed by discussions about Humbold’s ideal with its basic idea of the ‘unity of<br />

teaching and <strong>research</strong>’. Furthermore, there is a discussion of the ‘entrepreneurial<br />

university’, an institution which is no longer an ‘ivory tower’ but directly linked to the<br />

socio-economic demands of society and economy especially in the region where it is<br />

embedded. <strong>Universities</strong> fulfil various tasks, these include carrying out basic <strong>research</strong> and<br />

education, but more and more frequently also include applied <strong>research</strong> and development in<br />

co-operation with industry or public agencies, thereby trying to provide solutions <strong>for</strong> the<br />

immediate needs of society. These additional tasks are sometimes termed as a ‘third<br />

mission’ (OECD 1999).<br />

Moreover, the development of universities is influenced by the idea of New Public<br />

Management. The two major principles of New Public Management are the output and the<br />

per<strong>for</strong>mance orientation <strong>for</strong> public organisations, which, in turn, requires new budget<br />

allocation mechanisms and management <strong>systems</strong>. Management <strong>systems</strong> based on this idea<br />

have already been successfully implemented in many areas of the public sector.<br />

As organisations which are mainly financed by public funding, universities are also<br />

confronted with an increased demand by the owners and citizens <strong>for</strong> transparency<br />

regarding the use of those funds. This call <strong>for</strong> public accountability requires the disclosure<br />

about the social and economic outcomes of universities.<br />

Furthermore, universities are constantly faced with new paradigms <strong>for</strong> knowledge<br />

production, often labelled as ‘modus II of knowledge production’, which are characterised<br />

by a stronger orientation towards applied <strong>research</strong> and the necessity <strong>for</strong> an interdisciplinary<br />

<strong>research</strong> approach (Gibbons 1994). This implies the urge <strong>for</strong> universities to collaborate<br />

more frequently with other <strong>research</strong> institutions, but also with private firms or public<br />

organisations, or even to participate in international <strong>research</strong> networks. This development<br />

has been clearly accelerated by in<strong>for</strong>mation technologies, which enable more intensive cooperation<br />

and also provide new techniques <strong>for</strong> scientific analysing and modelling in an<br />

efficient way.<br />

<strong>Universities</strong> are part of the science, education, and innovation system of a nation and<br />

knowledge producers per se, their most important output is knowledge, incorporated in<br />

new <strong>research</strong> results, publications and well educated students. Obviously, the most<br />

valuable resources of universities are thus their <strong>research</strong>ers and students with their<br />

relationship networks as well as their organisational routines. These resources can be<br />

2


interpreted as intangible resources and assets, even though the term has so far not been<br />

used within the context of universities. Considering a wider definition of intangible assets -<br />

which is independent of the regulations by accounting standards’ committees and national<br />

laws, such as the right of disposal - as followed by most academics of the IC literature,<br />

intangible assets can doubtless be identified in universities.<br />

Due to their legal status, most universities do not have owner structures like private firms.<br />

Consequently, universities usually do not need to produce the kind of annual reports<br />

required by commercial law. Nevertheless, they have often to implement financial<br />

accounting <strong>systems</strong>, increasingly based on the double bookkeeping system as substitute <strong>for</strong><br />

the public accounting system.<br />

All the above mentioned challenges call <strong>for</strong> the implementation of new management and<br />

reporting <strong>systems</strong> within universities. In comparison with the past, universities are faced<br />

with the task of allocating budgets, they have to define their organisational goals and<br />

strategies in a more explicit way and due to the broader autonomy and the extended<br />

competition with other <strong>research</strong> organisations, universities gain a new leeway <strong>for</strong> setting<br />

strategic priorities. While in <strong>for</strong>mer times, when universities basically did only scientific<br />

<strong>research</strong> and education, today their <strong>research</strong> processes are much more manifold, ranging<br />

from basic <strong>research</strong> to the development of technologies. They educate full-time students<br />

but also offer training courses and provide labs <strong>for</strong> industrial applications. Nevertheless,<br />

university <strong>systems</strong> and traditions vary a great deal throughout Europe, a fact which has to<br />

be taken into account when dealing with the question of the management and measurement<br />

of the knowledge production process within universities.<br />

The instrument of intellectual capital (IC) reporting and the general methods <strong>for</strong> valuing<br />

intangible assets seem to be a rich foundation <strong>for</strong> the development of these new<br />

management and reporting <strong>systems</strong>. There<strong>for</strong>e in 2001 the Austrian Ministry <strong>for</strong><br />

Education, Science and Culture initiated a study of the potential of IC reporting. In 2002<br />

the Austrian Parliament finally decided that in the course of the reorganisation of Austrian<br />

universities, universities would have the obligation to publish IC Reports. In German IC<br />

reports are termed Wissensbilanz, which means knowledge account.<br />

The general conceptual framework and background <strong>for</strong> the implementation of IC reports<br />

will be first presented in this paper. Then this concept will be discussed in the context of<br />

other proposed management and valuation <strong>systems</strong> <strong>for</strong> universities. In universities the<br />

assessment of <strong>research</strong>, education, and knowledge-based processes has been mainly treated<br />

within the concepts of evaluation. In addition, the application of per<strong>for</strong>mance management<br />

<strong>systems</strong>, based on the principles of New Public Management, has been discussed and<br />

partly implemented. The paper compares the approach of evaluation, per<strong>for</strong>mance<br />

management and IC reporting and tries to point out the potential benefits and opportunities<br />

of IC reporting <strong>for</strong> universities. The differences between valuing intangible assets in<br />

universities, <strong>research</strong> organisation, and private firms will also be highlighted.<br />

3


2. A new management and reporting system <strong>for</strong> universities based on the<br />

principles of <strong>Intellectual</strong> <strong>Capital</strong> <strong>Reporting</strong><br />

2.1. The international development<br />

The concept of IC <strong>Reporting</strong> <strong>for</strong> universities is mostly based on experiences and methods<br />

developed within industry. The idea of IC reporting was born in the Scandinavian countries<br />

at the beginning of the nineties. Scandia, the Swedish Insurance Company, published the<br />

first IC report in 1992 as a supplement to its annual report. In the following years a new<br />

movement towards the disclosure of in<strong>for</strong>mation about the value and development of<br />

intangibles has attracted attention. In European and the US, companies have issued<br />

”accounts of intellectual capital“, which enable them to measure intangibles and their<br />

results. In the industrial application, approaches have been established which record<br />

intangible assets with the help of financial and non-financial indicators. Hereby different<br />

<strong>for</strong>ms of intangible assets are differentiated and each asset is valued with the help of<br />

indicators, <strong>for</strong> instance the length of product development, customer satisfaction, etc. (see<br />

<strong>for</strong> instance Sveiby 1997 and Edvinson 1997). Besides quantitative assessments methods<br />

such as qualitative valuations and narrations are also applied in these reports (Mouritsen et<br />

al. 1998).<br />

The idea of IC reporting did not gain much attention within the <strong>research</strong> and science sector.<br />

So far only Research Technology Organisations, which carry out applied <strong>research</strong> and<br />

development, have published <strong>Intellectual</strong> <strong>Capital</strong> Reports. In 1999 the Austrian Research<br />

Centers Seibersdorf published as the first European organisation an IC report <strong>for</strong> the entire<br />

organisation, followed by the German DLR, which published its first IC report in 2000.<br />

Both reports are based on a similar conceptual framework, developed within <strong>ARC</strong>, which<br />

allows at the same time the comparison of indicators between these two organisations. The<br />

Institute Optics and Fluid Dynamic of the Danish Risø <strong>research</strong> organisation has also had<br />

experiences with IC <strong>Reporting</strong> since 1999.<br />

In Germany, Maul (2000) dealt with the question of the exploratory power of financial<br />

statements, published by certain German universities in the course of the reorganisation in<br />

some German Federal States. He argues that the traditional financial account does not<br />

deliver sufficient in<strong>for</strong>mation on the financial situation, earnings and assets, which is the<br />

main task of the annual account according to the German HGB (Handelsgesetzbuch), the<br />

German accounting law. This is obviously due to the huge amount of intangible assets of<br />

universities. There<strong>for</strong>e, he claims that universities, which have to produce financial<br />

statements, should integrate in<strong>for</strong>mation about intangible assets within the management<br />

report of the financial statements.<br />

Based on the concepts and experiences with IC <strong>Reporting</strong> in general and with <strong>research</strong><br />

organisations in particular, a concept <strong>for</strong> the application of IC <strong>Reporting</strong> <strong>for</strong> Austrian<br />

universities was developed <strong>for</strong> the Austrian Ministry <strong>for</strong> Education, Science and Culture<br />

(see Leitner et al. 2001 and www.weltklasse-uni.at) 1 . The reorganisation of Austrian<br />

universities revealed a high demand <strong>for</strong> such a new concept and consequently a demand <strong>for</strong><br />

the implementation of new management and reporting <strong>systems</strong>.<br />

1 The <strong>research</strong> project was carried out by the <strong>ARC</strong> Seibersdorf <strong>research</strong> GmbH Systems Research and<br />

Montanuniversität Leoben, Institute <strong>for</strong> Economics and Business Management and funded the Austrian<br />

Ministry <strong>for</strong> Education, Science and Culture.<br />

4


2.2. The role of IC reporting within the Austrian reorganisation of universities<br />

The main task of the IC project was to develop a model <strong>for</strong> universities which meets the<br />

specifics of their knowledge production process in the new organisational and legal context<br />

of universities, <strong>for</strong>med by the national and European frameworks. The IC report also had to<br />

be integrated in the newly created administration and governance system of the university<br />

sector.<br />

The reorganisation of Austrian universities is based on the principles of New Public<br />

Management with the premises of increased autonomy, output orientation and<br />

per<strong>for</strong>mance-based funding (Titscher et al. 1999). Apart from new governance structures<br />

and labour laws, an important element to implement the trans<strong>for</strong>mation according to the<br />

new paradigm is the introduction of ‘per<strong>for</strong>mance contracts’ (Leistungsvereinbarungen).<br />

These contracts define the duties of both the universities (studies offered, human resources,<br />

<strong>research</strong> programs, co-operations and social goals) and the Ministry (funding), and assigns<br />

a global budget <strong>for</strong> the duration of three years. After three years new contracts are be<br />

drawn up. The funding, which is based on four criteria: need, demand, per<strong>for</strong>mance and<br />

social goals, is agreed between universities and the Ministry. It will be partly based on<br />

per<strong>for</strong>mance indicators up to 20% of the funds will be allocated, dependent on the<br />

development of selected per<strong>for</strong>mance measures, fixed within the per<strong>for</strong>mance contracts. In<br />

addition, every year the universities have to generate a per<strong>for</strong>mance report, which provides<br />

in<strong>for</strong>mation about the development and achievement of the contract.<br />

According to the new university law (see UOG 2002), the publication of the IC report has<br />

to be published parallel to the development of the per<strong>for</strong>mance contract and the<br />

per<strong>for</strong>mance report. While the per<strong>for</strong>mance report only deals with the topics addressed<br />

within the per<strong>for</strong>mance contract, the idea behind IC reports is to give universities the<br />

opportunity to report on their full range of activities and to provide an instrument <strong>for</strong><br />

internal management tasks and decision-making. The IC report has a wider scope than the<br />

per<strong>for</strong>mance contract and should allow the universities to in<strong>for</strong>m about their whole nature,<br />

unrestricted by a few measures. Both reports are generated by the university itself and it is<br />

the Vice Chancellor (Rektor) who is finally responsible <strong>for</strong> their proper implementation.<br />

IC reporting <strong>for</strong> Austrian universities should thus fulfil two aims: First, it should provide<br />

in<strong>for</strong>mation <strong>for</strong> the management of intangible resources. <strong>Universities</strong> have to decide<br />

whether, and how much they should invest in the training of scientists, with whom cooperations<br />

should be fostered, which <strong>research</strong> programs should be emphasised, etc. The<br />

implementation of an IC Report requires the discussion of goals and strategies, the IC<br />

model should support this discussion. With the IC report, which can be realised <strong>for</strong><br />

departments, institutes and the whole university, goals can be operationalised and their<br />

achievement monitored. Secondly, IC reports should provide external stakeholders with<br />

in<strong>for</strong>mation about the development and productive use of the intellectual capital. Thus, the<br />

Ministry gets useful in<strong>for</strong>mation <strong>for</strong> the allocation of its resources and the management of<br />

<strong>research</strong> programs but also <strong>for</strong> the definition of the national science and education policy.<br />

Besides the per<strong>for</strong>mance contract and IC reporting, evaluation is the third main instrument<br />

<strong>for</strong> the governance and development of Austrian universities. In Austria, evaluations do not<br />

have a long history and were only occasionally carried out in the past. In the course of the<br />

reorganisation, evaluations are to be institutionalised <strong>for</strong> all universities. They have a clear<br />

focus on in-depth assessments of the quality of <strong>research</strong> and education as well as the<br />

5


esearch personal and are to be carried out every three to five years by external peers (see<br />

also next chapter).<br />

Considering this background and guiding principles, the new university law finally defines:<br />

“The Wissensbilanz explicitly has to show i) the aim related to society, the self-defined<br />

strategy as well as the universities’ strategy, ii) the intangible assets, separated in Human<br />

<strong>Capital</strong>, Structural <strong>Capital</strong>, Relational <strong>Capital</strong>, iii) the per<strong>for</strong>mance processes and output<br />

measurements, in correspondence to the definition within the per<strong>for</strong>mance contract” (UOG<br />

2002). Every Austrian university has to produce the IC report annually by April 30 th , and<br />

report to the Ministry. It is up to the university to also publish it <strong>for</strong> other stakeholders.<br />

2.3. The IC model <strong>for</strong> Austrian universities<br />

The conceptual framework <strong>for</strong> IC reporting is based on a model which tries to trace the<br />

knowledge production process within universities (see Fig. 1). It consists of four main<br />

elements: the goals, the intellectual capital, the per<strong>for</strong>mance processes and the impacts.<br />

The model conceptualises the trans<strong>for</strong>mation process of intangible resources when carrying<br />

out different activities (<strong>research</strong>, education etc.) resulting in the production of different<br />

outputs according to the specific and general goals.<br />

Framework conditions<br />

Political<br />

goals<br />

Organisational<br />

goals<br />

<strong>Intellectual</strong><br />

capital<br />

Human capital<br />

Structural capital<br />

Relational capital<br />

Input<br />

Per<strong>for</strong>mance processes Impact<br />

Research<br />

Education<br />

Training<br />

Commercialising of <strong>research</strong><br />

Knowledge transfer to the public<br />

Services<br />

Infrastructure<br />

Fig. 1.: Model <strong>for</strong> IC <strong>Reporting</strong> of Austrian <strong>Universities</strong><br />

Source: Leitner et al. (2001)<br />

The development of intellectual capital is guided by political goals, set by the Ministry,<br />

which, in turn, are based on the Austrian Science and Education policy, but also by the<br />

organisational goals defined by the universities themselves. Within this model intellectual<br />

capital or intangible resources are interpreted as the input <strong>for</strong> the knowledge-production<br />

process within universities. Three elements of intellectual capital are differentiated: human<br />

capital, structural capital and relational capital. The model thus adopts the widely diffused<br />

classification also proposed by the MERITUM group (MERITUM 2001). Human <strong>Capital</strong><br />

are the <strong>research</strong>ers and non-scientific staff of the university. Structural capital are the<br />

routines and processes, an often neglected infrastructure of the universities. Relational<br />

capital are the relationships and networks of the <strong>research</strong>ers as well as the entire<br />

organisation.<br />

6<br />

Output<br />

Stakeholder:<br />

Ministry<br />

Students<br />

Industry<br />

Public<br />

Science<br />

Community<br />

etc.


The model differentiates six per<strong>for</strong>mance processes which can be reduced or enlarged<br />

during the implementation, depending on the specific profile and type of the university<br />

(e.g. universities <strong>for</strong> art, business schools). Apart from scientific <strong>research</strong> and education,<br />

which are the kernel of university activities, training, commercialising of <strong>research</strong>,<br />

knowledge transfer to the public, services, and infrastructure services are separated. The<br />

latter ones can be summarised as the third mission of the modern university (OECD 1999).<br />

These elements are mainly captured with process and output measures.<br />

In the category of impact, the achievements of the per<strong>for</strong>mance processes are assessed. The<br />

different stakeholders of the universities addressed are the scientific community, students,<br />

citizens, industry, etc. Naturally, these are the most difficult element to evaluate by<br />

quantitative data.<br />

As in the case of IC reporting by industrial firms the different elements of the model will<br />

be measured and assessed by financial and non-financial indicators, as well as by<br />

qualitative in<strong>for</strong>mation and valuations.<br />

Within the project a list of indicators was developed. Whereas some of them are obligatory<br />

indicators, every university has to publish, others are optional and can be used depending<br />

on the context and aims. The design and selection of indicators was based on i) the set of<br />

measures used in the past within Austrian universities, ii) proposed indicators within the<br />

intellectual capital literature, and iii) the findings of the evaluation <strong>research</strong>. A list of 200<br />

indicators was proposed, whereas 24 indicators will be obligatory <strong>for</strong> universities 2 . Even<br />

though most indicators are non-financial by nature, some indicators are financial figures.<br />

As mentioned, the financial assessment of outcomes is the most difficult one. In the case of<br />

commercialising, the amount of licensing and the sale of spin-out firms, are possible<br />

answers <strong>for</strong> that problem.<br />

Although scientific <strong>research</strong> and education are equally important <strong>for</strong> all universities,<br />

universities also have the choice to set certain priorities regarding their per<strong>for</strong>mance<br />

processes. This can be represented as the opposite levels of a continuum. On the one hand,<br />

<strong>for</strong> example, they can focus purely on basic <strong>research</strong>, or on the other hand carry out applied<br />

<strong>research</strong>. The latter is becoming increasingly important and can in turn deliver even new<br />

fundamental <strong>research</strong> questions. <strong>Universities</strong> can co-operate with industry or public<br />

agencies to solve concrete problems or commercialise their <strong>research</strong> findings via spin-offs .<br />

<strong>Universities</strong> may choose to educate only the best students and provide an elite education,<br />

while others provide education <strong>for</strong> masses. <strong>Universities</strong> will select the indicators according<br />

to their specific goals and the relevant processes. Thus, the list of indicators reflects the<br />

strategic priorities of a university.<br />

Finally, within the project a rough guideline <strong>for</strong> the implementation was also developed.<br />

The critical points <strong>for</strong> the implementation are the procedure and organisation of the<br />

implementation within the university (bottom-up versus top-down), the question of the<br />

participation of professors, administrators, <strong>research</strong> staff, etc., the co-ordination with other<br />

internal and external reporting <strong>systems</strong> as well as the successful use of the in<strong>for</strong>mation<br />

produced via the IC report and the establishment of feed-back loops. This implementation<br />

guideline will be further elaborated in 2003, when the new university law will be executed<br />

within Austrian universities. So far in Austria only the Institute <strong>for</strong> Economics and<br />

2 See Leitner et al. (2001) or http://www.weltklasse-uni.at/upload/attachments/170.pdf <strong>for</strong> the full list.<br />

7


Business Management from the Montanuniversität Leoben, which had been also a partner<br />

of the IC project and has a long experience in quality management, developed an IC report<br />

based on the presented IC model (Biedermann et al. 2002).<br />

3. <strong>Intellectual</strong> <strong>Capital</strong> reporting in the context of evaluation and<br />

per<strong>for</strong>mance management<br />

3.1. The evaluation of <strong>research</strong> and education<br />

Evaluations have the task to assess the effects, impacts and the efficiency of publicly<br />

funded institutions and programs. Evaluations started in the US in the 60s with the<br />

assessment of different public programs (regional development, employment, health etc.).<br />

They are linked to the long debate about how to assess public spending and how to<br />

estimate the social and economic effects of public programs and public <strong>research</strong><br />

expenditures. In the meantime, there is a considerable body that evaluates university<br />

<strong>research</strong> outcomes on the level of projects, programs, disciplines and institutions. In<br />

general, various approaches and concepts have been developed <strong>for</strong> the evaluation of the<br />

aims, scope and output of evaluations.<br />

While the US, the UK, and the Netherlands soon started to evaluate the <strong>research</strong> and<br />

education provided byuniversities, in other European countries the use of these concepts<br />

started quite late and the application and extent of evaluations still varies considerably<br />

across Europe. In most countries, evaluations are primarily carried out <strong>for</strong> institutes and<br />

departments or specific <strong>research</strong> programs, seldom <strong>for</strong> the entire university.<br />

In general, evaluations should deliver in<strong>for</strong>mation <strong>for</strong> public organisations and<br />

administration with the aim to solve or overcome identified problems and weaknesses.<br />

Evaluations are also seen as an instrument to promote re<strong>for</strong>ms and self-organisation within<br />

universities. However, in practice, they often fail to establish feed-back loops because,<br />

frequently, the reports are not discussed intensively across the evaluated institution<br />

(Blalock 1999).<br />

In addition to the purpose of delivering in<strong>for</strong>mation <strong>for</strong> the public authority, some authors<br />

argue that evaluations should also deliver in<strong>for</strong>mation <strong>for</strong> the general public and thus meet<br />

the demand <strong>for</strong> social accountability (Richter 1991).<br />

Finally, modern evaluations have the task of providing in<strong>for</strong>mation <strong>for</strong> the resource<br />

allocation. In practice, evaluations are partly related to resource allocation decisions. While<br />

in the UK the evaluations are directly associated with government resource allocations, in<br />

other countries such as the Netherlands evaluation reports are to a stronger extent aligned<br />

to support <strong>research</strong> groups and universities <strong>for</strong> their management tasks.<br />

Evaluations are based on different methods: quantitative tests, qualitative approaches,<br />

observations, bibliometry, self-assessments and benchmarking (Turpin et al. 1999).<br />

Evaluations are faced with methodological problems, especially regarding validity and<br />

credibility (Kieser et al. 1996). Despite the use of quantitative methods and data, most<br />

evaluation reports are still qualitative by nature. Because of their complex nature and<br />

diverse impacts, evaluations of public initiatives and programs are restricted to qualitative<br />

analyses.<br />

8


Evaluations are usually executed by external experts, in addition, internal evaluations have<br />

gained also popularity, especially because they are better suited to enhance learning within<br />

the organisation. The Dutch evaluation approach, <strong>for</strong> instance, starts with the preparation<br />

of the internal evaluation report, prepared by the institution (self-evaluation), followed by<br />

an evaluation by external experts (peer-group) (Richter 1991).<br />

When evaluating <strong>research</strong> and education, the goals have to be defined explicitly. However,<br />

goals of institutions and programs are often manifold, vague or even contradictory. A<br />

crucial task is thus the definition of valuation criteria, which are related to the goals and<br />

interest groups. Considering the specific aims, a major challenge <strong>for</strong> evaluations is the<br />

common definition of the output – which is especially difficult in the case of education and<br />

<strong>research</strong>, e.g. “What is a good education?” or “What are basic skills and learning<br />

capabilities?” This definition is, in turn, again dependent on the national and political<br />

context of education (see above).<br />

The problems discussed above also lead to the debate as to whether evaluations should<br />

generate recommendations. Rossi et al. (1988), <strong>for</strong> instance, argue that evaluation reports<br />

should only deliver data, whereas the assessment should be done by the contracting<br />

authority of the evaluation. Another variant is that the evaluator assesses on the basis of the<br />

goals that have been officially articulated. Shadish et al. (1991) proposed a “goal-freeevaluation”,<br />

which means that the evaluators have to work out the goals considering the<br />

context of the entire program or institution. Finally, constructivistic approaches argue that<br />

an objective specification of the conditions and an objective assessment, can never be<br />

achieved (Guba and Lincoln 1998). Even the evaluators have biased points of view and can<br />

thus never serve as objective authority.<br />

There is rich literature on evaluation of <strong>research</strong> and innovation theory, which deals with<br />

the scope of indicators. Important issues refer to the use of publications or patents, to<br />

measure <strong>research</strong> output or university-industry relationships (e.g. Dodgson and Hinze<br />

2000, Schartinger et al. 2000). The limitation of the use of the indicator ‘number of<br />

publications’ is widely recognised. The mere counting of publications is quite problematic<br />

since the quantity can hardly ever be equalled with the quality of publications and because<br />

occasionally quite subjective scientific paradigms and either the very conservative or very<br />

progressive attitudes of certain journals, can restrict or determine the publication of<br />

innovative results (Kieser et al. 1996).<br />

Regarding the organisation of and responsibility <strong>for</strong> evaluations, in many countries<br />

national agencies have developed guidelines <strong>for</strong> the preparation of evaluations. In the US<br />

and the UK, government legislation is the driving <strong>for</strong>ce <strong>for</strong> the systematic per<strong>for</strong>mance of<br />

evaluation and the definition of measures and indicators: GPRA in the US, Office of<br />

Science and Technology in the UK, whereas in the Netherlands the responsibility <strong>for</strong><br />

evaluation has been decentralised and delegated to intermediary organisations, such as<br />

funding foundations or university associations. The above mentioned Dutch model focuses<br />

strongly on self-assessment, combined with a peer review approach.<br />

3.2. The concept of per<strong>for</strong>mance management<br />

In the 80s, especially in the Anglo-American countries, public organisations such as<br />

hospitals, schools or public agencies started to implement per<strong>for</strong>mance management<br />

<strong>systems</strong>, which are rooted in the idea of New Public Management, as well as the adoption<br />

9


of ideas and methods successfully used in the private sector. Some proponents even<br />

explicitly state that they desire to make the public sector more like the private sector with<br />

respect to accountability (Itner and Larcker 1998). Herein, the idea of output and outcome<br />

orientation is central. In terms of classical accounting, per<strong>for</strong>mance management can be<br />

classified as a “management control system” (Cunningham 2000).<br />

Per<strong>for</strong>mance management is clearly a management and planning tool. Like evaluations,<br />

per<strong>for</strong>mance management <strong>systems</strong> try to base judgements of the effectiveness of social<br />

programs and public institutions on more appropriate and trustworthy in<strong>for</strong>mation,<br />

however, the approach is different.<br />

Per<strong>for</strong>mance management is the process of the definition of goals and per<strong>for</strong>mance<br />

categories, the definition of per<strong>for</strong>mance standards and the reporting and communication to<br />

the public owners (National Academy of Public Administration, cited in English and<br />

Linquist 1998). Thereby, initial objectives <strong>for</strong> programs and institutions have to be defined,<br />

<strong>for</strong> which, in turn, relavantindicators have to be developed. This process allows to assess<br />

the achievement of objectives. A crucial activity in designing per<strong>for</strong>mance management<br />

<strong>systems</strong> is the development of per<strong>for</strong>mance measures, which include measures <strong>for</strong> outputs<br />

and outcomes, sometimes also <strong>for</strong> inputs and processes. This leads to the necessity to<br />

define the outputs as well as the products. As discussed above, the definition and<br />

measurement of outputs of universities is provocative. If the inputs and outputs are<br />

quantified by indicators, it is possible to link inputs and outputs and to develop efficiency<br />

measures (Davies 1999). In addition, with these indicators the allocation of resources can<br />

be carried out. In some European countries universities and Ministries started to use<br />

contracts to allocate the funds in accordance with per<strong>for</strong>mance indicators, a method which<br />

is in line with the idea of per<strong>for</strong>mance management (Ziegele 2001).<br />

Basically, per<strong>for</strong>mance management <strong>systems</strong> use financial as well as non-financial<br />

indicators with the main aim to define the outcome and they collect quantitative measures,<br />

often based on IT-based <strong>systems</strong> or Management In<strong>for</strong>mation Systems. In practice,<br />

however, they often produce simple statistics to answer critical planning and managerial<br />

questions, instead of conducting more complex data analyses. Thus, they are often<br />

criticised as using quick-turn-around data in internal automated Management In<strong>for</strong>mation<br />

Systems that delivers relatively simple comparisons of gross outcomes against<br />

per<strong>for</strong>mance goals (Blalock 1999).<br />

Though criticised that the true complexity of programs and institutions is not captured by<br />

per<strong>for</strong>mance management <strong>systems</strong>, proponents argue that these <strong>systems</strong> can free agencies<br />

and staff from rigid regulations (Cunningham 2000). Furthermore, the definition of<br />

per<strong>for</strong>mance measures is consistent with program goals or institutional goals and supports<br />

hence the strategic development process at all levels of an institution. Hereby one<br />

important challenge <strong>for</strong> public organisations is the evolution of the use of outcome instead<br />

of output indicators (Cunningham 2000).<br />

A widely recognised problem with per<strong>for</strong>mance management is the ‘goal displacement’,<br />

which occurs when per<strong>for</strong>mance management creates incentives that direct ef<strong>for</strong>t towards<br />

meeting the requirements of measuring and reporting and not towards the overall aims of<br />

the institution (Davies 1999). This leads to a behaviour where people will work “first <strong>for</strong><br />

the indicators”. Empirical studies also stress the problem of “bean counting“ in practice<br />

(Cunningham 2000). Obviously, this is a problem <strong>for</strong> quantitative evaluations, too.<br />

10


Critical factors <strong>for</strong> a successful implementation are the top leadership support, the personal<br />

involvement of the senior management, the participation of the relevant stakeholders and<br />

clarity about how the per<strong>for</strong>mance in<strong>for</strong>mation shall be used (Greene 1999). Cunningham<br />

(2000) found that the commitment and involvement of all actors as well as the integration<br />

in the strategic planning process is critical. The benefits arise rather from the process itself<br />

than from the end result of the per<strong>for</strong>mance reporting activity. In particular, the<br />

implementation of per<strong>for</strong>mance management <strong>systems</strong> requires communication about<br />

per<strong>for</strong>mance areas that have not been present be<strong>for</strong>e. Thus, it is rather the communication<br />

that leads to benefits than the per<strong>for</strong>mance in<strong>for</strong>mation itself.<br />

3.3. IC reporting, evaluation and per<strong>for</strong>mance management at a glance<br />

The three instruments presented have different aims and scopes of applications, which<br />

clearly overlap to some extent. In the following the three instruments will be discussed<br />

according to their different features and aims (see Fig. 2 <strong>for</strong> an overview).<br />

Regarding the approach and focus of the instruments, evaluation is the scientific-oriented<br />

approach with in-depth analysis, carried out at certain points of time, while per<strong>for</strong>mance<br />

management <strong>systems</strong> are characterised by a pragmatic approach, primarily based on output<br />

indicators, derived from the goals. In practice, both are often implemented simultaneous.<br />

IC reporting <strong>for</strong> universities, as presented in this paper, is a tool which encloses the entire<br />

knowledge production process within universities, with the aim of generating in<strong>for</strong>mation<br />

<strong>for</strong> management decisions. It is mainly based on indicators which are categorised in a<br />

model. In addition, it also incorporates qualitative in<strong>for</strong>mation, that especially should<br />

express the complex nature and interdependencies between driving factors and results.<br />

In contrast to per<strong>for</strong>mance management <strong>systems</strong> and evaluation, IC reports focus explicitly<br />

on the intellectual capital and hence enlarge the existing input and output categories of<br />

per<strong>for</strong>mance management <strong>systems</strong>. Particularly structural capital has to be considered as a<br />

blind spot within universities. Based on the model presented in this paper the importance<br />

of intangible resources is highlighted <strong>for</strong> universities. IC reporting relies on quantitative as<br />

well as qualitative valuations, whereas per<strong>for</strong>mance management <strong>systems</strong> often focus on a<br />

few indicators.<br />

All three instruments apply different kinds of qualitative and quantitative methods.<br />

Although evaluations rely more often on qualitative methods, in recent years these have<br />

increasingly integrated different kinds of indicators. Indicators have already become<br />

sometimes an integral part or method of certain evaluations, especially in the US, where<br />

data about the financial resources, student satisfaction, or academic reputation, partly<br />

gathered by questionnaires, is aggregated via a standardised procedure (Kellermann 1992).<br />

The classifications of input, processes and results is commonly used in all three<br />

instruments. Stufflebeam (1983), <strong>for</strong> instance, proposes that evaluations have to analyse<br />

the context, e.g. what are the aims addressed with the program?, inputs, e.g. human<br />

resources and tangible resources, process, activities by the program or institutions, and<br />

products, the results of the program. This is also applied by the IC model presented in this<br />

paper.<br />

11


focus<br />

approach /<br />

methods<br />

per<strong>for</strong>mance<br />

management<br />

measuring results via<br />

indicators<br />

per<strong>for</strong>mance indicators ,<br />

gathered via in<strong>for</strong>mation<br />

<strong>systems</strong><br />

mainly based on<br />

in<strong>for</strong>mation <strong>systems</strong> &<br />

databases<br />

data collection Continuous<br />

based on in<strong>for</strong>mation<br />

support shareholder/<br />

public authority<br />

internal management<br />

support<br />

resource allocation<br />

recommendation<br />

responsibility<br />

<strong>systems</strong><br />

12<br />

external evaluation intellectual capital<br />

assessment of efficiency<br />

and effectiveness<br />

scientific-oriented and indepth<br />

tailored to the specific<br />

goals<br />

various qualitative and<br />

quantitative<br />

often qualitative<br />

report<br />

periodically (3-5 years) or<br />

at discrete points<br />

reporting 3<br />

intangible resources and<br />

its management<br />

indicator-based<br />

obligatory indicators, but<br />

flexible <strong>for</strong> adaptation<br />

indicators<br />

narrations<br />

report<br />

periodically,<br />

preferable based on<br />

in<strong>for</strong>mation <strong>systems</strong><br />

partly major aim in<strong>for</strong>mation <strong>for</strong> policy<br />

decisions<br />

deliver in<strong>for</strong>mation <strong>for</strong><br />

evaluations<br />

restricted deliver in<strong>for</strong>mation <strong>for</strong><br />

evaluations<br />

yes, based on indicators yes, mainly in the US possible<br />

no partly no<br />

organisation (unit) national agencies organisation (unit)<br />

Fig. 2.: Comparison of evaluation, per<strong>for</strong>mance management and IC reporting<br />

With respect to the measurement of results, the separation into output, outcome and<br />

impact, whereby the latter two are sometimes used synonymously, is used frequently in<br />

evaluations (Garrett-Jones 2000). In general, output refers to the routine products of<br />

<strong>research</strong> activities, such as publications, conference papers, training courses, degrees etc.<br />

Outcome means the achievements of the activity such as new theories, new devices or<br />

analytical techniques. Outcome is not treated separately in the IC model. Impact is a<br />

measure of the influence or benefit of the <strong>research</strong> outcome or output, either within the<br />

<strong>research</strong> community itself or the wider society. It thus includes the social, economic or<br />

environmental benefit. The assessment of impacts is explicitly treated in the IC model.<br />

The aspects of learning and supporting management is addressed differently by the three<br />

instruments. Per<strong>for</strong>mance management <strong>systems</strong> document outcomes, but to a lesser extent<br />

give in<strong>for</strong>mation to understand the complex nature of knowledge-production. They collect<br />

data on a continuous basis, which tends to rely on easy-to-obtain quantitative measures that<br />

are conditioned by cost acceptance. In the field of evaluations, internal evaluations have a<br />

strong focus to increase the per<strong>for</strong>mance and efficiency by internal discussions. IC reports<br />

should also enhance the academic discussion within the faculty board (dean) and the<br />

university’s governing boards and should there<strong>for</strong>e support the strategic development<br />

process.<br />

The linking of per<strong>for</strong>mance measurement to budgeting is a global phenomenon and is<br />

frequently treated within per<strong>for</strong>mance management <strong>systems</strong>. IC reports have the potential<br />

3 As conceptualised <strong>for</strong> the implementation in Austrian universities.


to support this task. However, in Austria, the allocation of financial resources will be based<br />

on the per<strong>for</strong>mance contract (see above). Within this procedure, the IC report supports the<br />

management and decision process, either <strong>for</strong> the Ministry, or <strong>for</strong> the internal resource<br />

allocation decisions between university departments, <strong>research</strong> programs, etc. In such a<br />

case, data also have to be indicated on the department level. However, the latter is not<br />

demanded by the new university law and is the responsibility of the universities<br />

themselves.<br />

As concerns the responsibility <strong>for</strong> accomplishing the instruments, traditional evaluations<br />

are carried out by an external expert team, which, based on different methods, generates<br />

data, interprets it and writes a report. Internal evaluations are produced by the members of<br />

the organisation itself. Per<strong>for</strong>mance management is implemented within the organisation<br />

and administrated by the top management or administration. IC reports are produced by the<br />

members of the organisation, as is the case in industry. Probably, in the future, external<br />

authorities will control the process and ensure the quality of the production, <strong>for</strong> instance<br />

auditors or Ministry agents.<br />

4. Discussion and Outlook<br />

Considering the long history of assessing results of <strong>research</strong> and education, the idea of<br />

measuring and reporting knowledge-based resources and results <strong>for</strong> application at<br />

universities is not really brand new. However, IC reporting as conceptualised here focuses<br />

on the identification of intangible assets and tries to link them to the outcome of the<br />

universities, which is a new idea in the context of universities. The preparation of an IC<br />

report requires the discussion and definition of a university’s aims and strategies, which, in<br />

turn, stimulates learning inside the organisation. IC reports, regularly published, and based<br />

on in<strong>for</strong>mation <strong>systems</strong>, deliver in<strong>for</strong>mation <strong>for</strong> monitoring the development of the<br />

organisation and supports the organisation university to develop a coherent strategic<br />

profile. It fosters the establishment of an organisational structure and culture within<br />

universities, which is increasingly important <strong>for</strong> many universities in the new competitive<br />

environment. When taking intangibles explicitly into account, the efficient use of resources<br />

can be improved. However, aims, rationales and the knowledge production processes of<br />

universities are considerably different in comparison to private industry. This phenomenon<br />

has to be reflected by the specific indicators and aims incorporated in the IC model and<br />

reports of universities.<br />

Like knowledge-based enterprises, IC reporting universities are faced with the<br />

methodological problem of measuring the ‘soft’, non-physical processes and outputs.<br />

There is a long history and vast literature dealing with the measurement of scientific and<br />

educational processes and their per<strong>for</strong>mance by indicators within the evaluation <strong>research</strong><br />

and innovation theory. A common problem is that indicators may lack validity and may not<br />

be the best proxy <strong>for</strong> outcome variables. In general, as one moves from output to impacts,<br />

the results are getting broader in their effect, need a period to manifest, are harder to<br />

quantify and less traceable to a particular <strong>research</strong> project, program or organisation<br />

(Garrett-Jones 2000). The definition of the output of universities – a public good - is, as<br />

already described, difficult. Due to the limitations of a purely indicator-based method, IC<br />

reports should also integrate qualitative methods (best practice, narration, etc.), <strong>for</strong><br />

instance, to illustrate the connection between <strong>research</strong> and education and to show the<br />

synergies and flows between the different elements of the model. However, currently there<br />

13


are no quantitative measures available in most IC approaches, which would allow to<br />

monitor the flows or synergies between the different elements.<br />

In general, in the field of <strong>research</strong> there are still concerns within the community regarding<br />

the measurement of <strong>research</strong> outputs and the establishment of quantitative or financial<br />

methods which goes hand in hand with the fear that curiosity, discovery, creativity, and<br />

innovation will be restricted too much. However, this might be the wrong perception, it is<br />

not about restraining the process itself, but taking decisions on scarce resources in an<br />

increasingly complex environment. IC reporting should serve as an instrument to take<br />

more strategic, efficient and transparent decisions, taking into account different<br />

perspectives simultaneously. This debate has also similarities with the discussion within<br />

the knowledge management literature and the question, as to whether or not one can<br />

manage knowledge at all.<br />

In industry there is a strong demand <strong>for</strong> the valuation of intangible assets by financial<br />

figures, with mixed and successful solutions offered. In the case of universities, this call is<br />

not that loud, but existent. The financial assessment of results is related to the estimation of<br />

the economic effects of <strong>research</strong> and education. Market approaches are seldom applicable<br />

<strong>for</strong> universities. Nevertheless, some indicators, which are based on questionnaires,<br />

elevating in<strong>for</strong>mation such as the salary of <strong>for</strong>mer students, can be seen as such financial<br />

measures. Based on the employment or profits generated with these firms, an economic<br />

benefit might also be estimated. However, the development is still in the early stages and<br />

has to be treated with caution within the universities.<br />

In general, the valuation of IC indicators is dependent on the specific goals and the context<br />

of the university or institute. This issue is also widely stressed in the IC literature (e.g.<br />

Roberts 1999). IC reports can be developed <strong>for</strong> the entire university, <strong>for</strong> institutes,<br />

departments and <strong>research</strong> programs. Clearly, this causes additional problems of<br />

aggregation and comparison. An interesting question, <strong>for</strong> instance is the possibility of<br />

valuing and comparing the per<strong>for</strong>mance of a <strong>research</strong>-oriented institute with a more<br />

training-oriented institute. In practice, the interpretation of the reports is dependent on the<br />

context. This becomes especially evident <strong>for</strong> education, since the meaning and value of<br />

education is contingent on the cultural context. In the German Humboldtian tradition, <strong>for</strong><br />

instance, education is referred to as “scientific”, <strong>for</strong> the French, education is regardedas<br />

“professional”, which also explains France’s phenomenon of the “grande ecoles”, finally,<br />

the Anglo-American tradition stresses the “liberal” nature of education (Scott 2002). These<br />

few examples already point out the necessity <strong>for</strong> IC reports to deliver contextual<br />

in<strong>for</strong>mation, but also shows the limits of comparability. This is thus a further complicating<br />

aspect of IC reports of universities in contrast to IC reports of industrial firms. While in the<br />

latter case, eventually all measures can be linked (explicitly) to the financial per<strong>for</strong>mance<br />

of the firm, in the <strong>for</strong>mer, the per<strong>for</strong>mance measurement is per se multi-dimensional,<br />

expressing different aims and objectives.<br />

If IC reports serve as a management instrument, the generated in<strong>for</strong>mation has to be linked<br />

to action and behaviour and hence the establishment of learning loops is crucial. In<br />

comparison to IC reporting in industry, the presented IC model has a stronger focus on the<br />

management perspective. The general lack of accounting, management and control<br />

instruments in universities has thus to be keep in mind in this context. Empirical studies,<br />

carried out in industry regarding the introduction of the Balanced Score Sard, delivered<br />

some evidence that indicator <strong>systems</strong> helped the firms to make their corporate goals and<br />

14


strategies more concrete and measurable, which might be also come true in the context of<br />

universities (Hoque and James 1999).<br />

Obviously, some problems might emerge with the two aims, namely to serve as a<br />

management instrument <strong>for</strong> the university itself and also <strong>for</strong> the ministry. Institutes will not<br />

be willing to deliver in<strong>for</strong>mation and especially intimate in<strong>for</strong>mation, if they have the fear<br />

that this will have negative consequences <strong>for</strong> their funding. The problem of tactical<br />

behaviour due to evaluations is well recognised in the corresponding literature, especially<br />

if results are bound to financial allocation mechanisms (Voroeijentijen 1991, Blalock<br />

1999).<br />

As far as the structure of the IC model applied to the Austrian universities is concerned, it<br />

is different to most classical IC models <strong>for</strong> industry since it follows a process logic or<br />

input-output logic and should thus be labelled as ‘process-oriented model’. Classical<br />

models of IC <strong>Reporting</strong> separate different elements of intellectual capital but seldom link<br />

this with the production process of the organisation. An exception are IC reports, based on<br />

EFQM model such as the Danish company Ramboll applies <strong>for</strong> its IC measurement<br />

(Ramboll 2001). In current models the development of the intangible resources is explicitly<br />

embedded in the context of goals, processes and outputs. However, like most other<br />

approaches <strong>for</strong> IC measurement, the presented model is not directly compatible or cannot<br />

be incorporated in the classical financial accounting system.<br />

As described above, evaluation, per<strong>for</strong>mance management and IC reporting are<br />

overlapping to some extent, but offer different solutions <strong>for</strong> the same problems. The latter<br />

are partly substitutes <strong>for</strong> the <strong>for</strong>mer because the organisation and its per<strong>for</strong>mance can be<br />

analysed by internal and external stakeholders without the help of the evaluators. The<br />

question of the balance of internal self-assessment versus external evaluation is a crucial<br />

one <strong>for</strong> the further development and IC reporting has to be linked to this question. Within<br />

this context, universities and ministries will have to develop their <strong>systems</strong> <strong>for</strong> governance,<br />

management, and control, considering the integration ef<strong>for</strong>ts within Europe. But there is no<br />

doubt, with the variety of in<strong>for</strong>mation provided <strong>for</strong> the stakeholders, IC reporting allows<br />

the comparability between different universities and enables quality assurance at<br />

universities, something which is demanded in the course of the harmonisation of European<br />

university <strong>systems</strong> with its principles of transparency, per<strong>for</strong>mance and quality assessment<br />

and an open dialogue with the community.<br />

What are possible development trajectories <strong>for</strong> the future? The presented functional model<br />

<strong>for</strong> application in Austrian universities is a first step. It is flexible enough <strong>for</strong> individual<br />

adaptations and adjustments and has the potential <strong>for</strong> further improvement and elaboration.<br />

Furthermore, through the definition of categories and defined indicators, comparability<br />

should be made possible. Although the various national university <strong>systems</strong> still vary<br />

considerably, this instrument, which is flexible <strong>for</strong> modification and improvement,<br />

addresses some general issues relevant <strong>for</strong> all universities, independent of their national<br />

specifics. On the whole, there is so far no common model or standard <strong>for</strong> valuing<br />

intangible assets and preparing IC reports <strong>for</strong> universities at the international level. In<br />

many countries, national agencies have developed guidelines <strong>for</strong> the preparation of<br />

evaluations and have partly also defined some indicators, which have to be linked to and<br />

co-ordinated with the further development within the IC movement at universities. In<br />

addition, work carried out at the national and international level regarding IC management<br />

and measurement provide valuable in<strong>for</strong>mation <strong>for</strong> the development of specific guidelines<br />

15


<strong>for</strong> the application within universities (e.g. MERITUM <strong>for</strong> HEROs – Higher Education and<br />

Research Organisations, currently in development).<br />

There is also the demand <strong>for</strong> further investigation of the synergies of the different<br />

approaches and to link the three disciplines: evaluation of <strong>research</strong>, per<strong>for</strong>mance<br />

measurement and IC theory. Furthermore, universities can not only learn from the private<br />

sector and IC theory, private <strong>research</strong> organisations and not-<strong>for</strong> profit <strong>research</strong><br />

organisations as well as R&D departments of companies, can also benefit from the<br />

experiences of universities with the management of <strong>research</strong> and human capital.<br />

References<br />

Austrian Research Centers Seibersdorf (2000): <strong>Intellectual</strong> <strong>Capital</strong> Report 1999.<br />

Seibersdorf 2000.<br />

Biedermann, H., Graggober, M., Sammer, M. (2002): Die Wissensbilanz als Instrument<br />

zur Steuerung von Schwerpunktbereichen am Beispiel eines Universitätsinstitutes. In:<br />

Wissensmanagement: Konzepte und Erfahrungsberichte aus der betrieblichen Praxis.<br />

Bornemann, M., Sammer, M. (Edt.), Wiesbaden 2002 S. 53-72.<br />

Blalock, A.B. (1999): Evaluation Research and the Per<strong>for</strong>mance Management Movement,<br />

Evaluation, 5, 2, 117-149.<br />

Bundesministerium für Bildung, Wissenschaft und Kultur - BM:BWK (2002):<br />

Bundesgesetz über die Organisation der Universitäten und ihrer Studien<br />

(Universitätsgesetz 2002), Wien: URL: http://www.weltklasseuni.at/upload/attachments/150.pdf<br />

Bomaccorsi, A., Piccaluga, A. (1994): A Theoretical Framework <strong>for</strong> The Evaluation of<br />

University - Industry Relationships, R&D Management, 24, 3, 154-169.<br />

Bormans, M.J, Brouwer, R. In’t Veld, R.J., Merten, F.J. (1987): The role of per<strong>for</strong>mance<br />

indicators in improving the dialogue between government and universities, in:<br />

International Journal of Institutional Management in Higher Education, 11, 181-194.<br />

Broadbent, J., Guthrie, J. (1992): Changes in the Public sector: A review of recent<br />

“alternative” accounting <strong>research</strong>, Accounting, Auditing & Accountability Journal, 5, 2 3-<br />

31.<br />

Canibano, L./Covarsi, M./Sanchez M.P. (1999): The value relevance and managerial<br />

implications of intangibles: A literature review, International Symposium: Measuring and<br />

<strong>Reporting</strong> <strong>Intellectual</strong> <strong>Capital</strong>: Experiences, Issues, and Prospects, OECD, 9-10 June,<br />

Amsterdam.<br />

Conceicao, P., Heitor, M. (1999): On the Role of the University in the Knowledge<br />

Economy. Science and Public Policy, 26, 1, 37-51.<br />

Cooper, R. et al. (1981): Accounting in organised anarchies: Understanding and designing<br />

accounting system in ambiguous situations, Accounting, Organizations and Society, 6, 3,<br />

175-191.<br />

16


Cronbach, L.J., Ambron, S., Dornbusch, S.M., Hess, S.M., Hornik, R.C. et al. (1980):<br />

Towards Re<strong>for</strong>m of Program Evaluation, San Francisco.<br />

Cunningham, G. (2000): Towards A Theory of Per<strong>for</strong>mance <strong>Reporting</strong> in Achieving<br />

Public Sector Accountability: A Field Study. Paper presented at the Annual Meeting of the<br />

British Accounting Association 2000.<br />

Davies, I.C. (1999): Evaluation and Per<strong>for</strong>mance Management in Government. In:<br />

Evaluation, 5, 5, 150-159.<br />

Deutsches Zentrum für Luft- und Raumfahrt DLR (2001): <strong>Intellectual</strong> <strong>Capital</strong> Report 2000,<br />

Köln 2001.<br />

Dodgson, M., Hinze, S. (2000): Indicators used to measure the innovation process: defects<br />

and possible remedies, Research Evaluation, 8, 2, 101-114.<br />

Edvinsson, L. (1997): <strong>Intellectual</strong> <strong>Capital</strong>, Skandia, Stockholm 1997.<br />

English, J., Lindquist, E. (1998): Per<strong>for</strong>mance Management: Linking Results to Public<br />

Debate, Ottawa.<br />

Etzkowitz, H., Webster, A., Heavley, I. (1998): <strong>Capital</strong>izing Knowledge: New<br />

Intersections of Industry and Academia, State University of New York Press, Albany.<br />

Etzkowitz, H., Leydesdorff, L. (2000): The dynamics of innovation: from National<br />

Systems and “Mode 2” to a Triple Helix of university-industry-government relation,<br />

Research Policy, 29, 109-123.<br />

European Foundation <strong>for</strong> Quality Management: URL: http://www.efqm.org/<br />

Garrett-Jones, S. (2000): International trends in evaluation university <strong>research</strong> outcomes:<br />

what lessons <strong>for</strong> Australia.<br />

Gibbons, M. (1994): The New Production of Knowledge, Pinter Publishers, London and<br />

New York.<br />

Ginsberg, P.E. (1984): The Dysfunctional Side of Quantitative Indicator Production. In:<br />

Evaluation and Program Planning, 7, 1-12.<br />

Guba, E.G., Lincoln, Y.S. (1989): Fourth generation evaluation, Newbury Parc, CA.<br />

Hoque, Z., James, W. (1999): Strategic Priorities, Balanced Scorecard Measures and their<br />

Interaction with Organizational Effectiveness: An Empirical Investigation, British<br />

Accounting Association, Annual Conference, March 29-31 1999, University of Glasgow.<br />

Itner, C.D., Larcker, D.F. (1988): Innovations in per<strong>for</strong>mance measurement: Trends and<br />

<strong>research</strong> implications, Journal of Management Accounting Research, 10, 205-238.<br />

17


Jaffe, A. (1989): Real effects of academic <strong>research</strong>. In: American Economic Review 79. Jg<br />

(1989), S. 957-970.<br />

Kieser, A. (1998): Going Dutch – Was lehren niederländische Erfahrungen mit der<br />

Evaluation universitärer Forschung. In: Die Betriebswirtschaft, 58, 2, 208-224.<br />

Leitner, K.-H., Sammer, M., Graggober, M., Schartinger, D., Zielowski, C. (2001):<br />

Wissensbilanzierung für Universitäten, Auftragsprojekt für das Bundesministerium für<br />

Bildung, Wissenschaft und Kunst. 2001. URL: http://www.weltklasse-uni.ac.at.<br />

Lenn, M.P. (1992): The US accreditation system, in: Craft, A. (Edt.): Quality Assurance in<br />

Higher Education. Proceedings of an International Conference Hong Kong, 1991, London,<br />

161-186.<br />

Lindgren, L. (2001): The Non-profit Sector Meets the Per<strong>for</strong>mance-management<br />

Movement. In: Evaluation, 7, 3, 285-303.<br />

Lundvall, B. (2002): The University in the Learning Economy, DRUID Working Paper No<br />

02-06, Copenhagen<br />

Maul, K-H. (2000): Wissensbilanzen als Teil des handelsrechtlichen Jahresabschlusses. In:<br />

Deutsches Steuerrecht, 38, 4, 2009-2016.<br />

MERITUM Project (2001): Guidelines <strong>for</strong> Managing and <strong>Reporting</strong> on Intangibles<br />

(<strong>Intellectual</strong> <strong>Capital</strong> Report).<br />

Mouritsen, J., Larsen, H.T., Bukh, P.N.D. (1998): <strong>Intellectual</strong> <strong>Capital</strong> and the ‘Capable<br />

Firm’: Narrating, Visualising and Numbering <strong>for</strong> Managing Knowledge, Copenhagen<br />

Business School and Aarhus School of Business 1998.<br />

National Science Board (1998): Science and Engineering Indicators - 1998, National<br />

Science Foundation, Arlington, VA.<br />

OECD (1999): University Research in Transition, Paris 1999.<br />

Ramboll (2001): Annual Report 2000, Copenhagen.<br />

Richter, R. (1991): Qualitätsevaluation von Lehre und Forschung an den Universitäten der<br />

Niederlande. Eine Bilanz der letzten 10 Jahre, in: Weber, W. Otto, H. (Edt.): Der Ort der<br />

Leher in der Hochschule, Weinheim, 337-362.<br />

Rossi, P., Freeman, H.E., Hofmann, G. (1988): Programm-Evaluation. Einführung in die<br />

Methoden angewandter Sozial<strong>for</strong>schung, Stuttgart.<br />

Risø National Laboratory Optics and Fluid Dynamic Department (1999): <strong>Intellectual</strong><br />

<strong>Capital</strong> Accounts, Roskilde.<br />

Roberts, H. (1999): The Control of Intangibles in the Knowledge-intensive Firm. Paper<br />

presented at the 22 nd Annual Conference of the European Accounting Association,<br />

Bordeaux, 1999.<br />

18


Rosenberg, N., Nelson, R.R. (1996): The Roles of <strong>Universities</strong> in the Advance of Industrial<br />

Technology. In: Rosenbloom, R.S. and W.J. Spencer (Edt.): Engines of Innovation.<br />

Cambridge: Harvard Business School Press.<br />

Salamanca (2001): The European Higher Education Area. Joint declaration of the<br />

European Ministers of Education. Convened in Bologna on the 19 th of June 1999.<br />

http://www.salamanca2001.org/<br />

Scott, P. (2002). The future of general education in mass higher education <strong>systems</strong>, Higher<br />

Education Policy 15, 61-75.<br />

Scriven, M. (1993): Hard-won lessons in program evaluation, San Francisco, CA<br />

Schartinger D., Schibany A., Gassler H. (2001): Evidence of Interactive Relations Between<br />

the Academic Sector and Industry. In: The Journal of Technology Transfer, 26, 3.<br />

Schenker-Wicki, A. (1996): Evaluation von Hochschulleistungen, Leistungsindikatoren<br />

und Per<strong>for</strong>mance Measurement. Wiesbaden 1996.<br />

Schimanek, U., Winnes, M. (2000): Beyond Humboldt? The relationship between teaching<br />

and <strong>research</strong> in European university <strong>systems</strong>. In: Science and Public Policy, 27, 6, 397-<br />

408.<br />

Stufflebeam, D.L. (1983): The CIPP model <strong>for</strong> program evaluation, in: Madaus, G.F.,<br />

Scriven, M.S., Suffelbeam, D.L. (Edt.): Evaluation models. Viewpoints on Education and<br />

Human Service Evaluation, Boston, 117-142.<br />

Stephan, P. E. (1996): The Economics of Science, Journal of Economic Literature, 34. Jg.<br />

(1996), S. 1199-1235.<br />

Sveiby, K.E. (1997): The New Organizational Wealth: Managing and Measuring<br />

Knowledge-Based Assets, Berrett-Koehler, San Francisco.<br />

Titscher, S. et al. (Edt.) (2000): Universitäten im Wettbewerb. München 2000.<br />

Turpin, T., Garrett-Jones, S., Aylward, D. et al. (1999): Valuing University Research:<br />

International Experiences in Monitoring and Evaluation Research Outputs and Outcomes.<br />

Australian Research Council and Centre <strong>for</strong> Research Policy, University of Wollongong,<br />

May 1999.<br />

Vroeijenstijn, T. (1992): External quality assessment: Servant of two masters? The<br />

Netherlands university perspective, in Craft, A. (Edt.): Quality Assurance in Higher<br />

Education: Proceedings of an International Conference Hong Kong, 1991, London, 109-<br />

132.<br />

Ziegele, F. (2001): Akademisches Controlling: Theoretische Grundlagen, Ziele, Inhalte<br />

und Ergebnisse. In: Akademisches Controlling und hochschulinterne Zielvereinbarung,<br />

Kooperationsprojekt der Technischen Universität München und des CHE Centrum für<br />

Hochschulentwicklung, München, Gütersloh Jänner 2001.<br />

19


Appendix: Exemplary indicators <strong>for</strong> Austrian universities<br />

(obligatory indicators are labelled with an asterisk)<br />

Human <strong>Capital</strong><br />

Number of scientific staff total*<br />

Number of scientific staff total (employed)*<br />

Number of full-time professors<br />

Number of student assistants<br />

Fluctuation of scientific staff (as % of all scientific staff)*<br />

Fluctuation of scientific staff (not empl.) (as % of total scientific staff (not empl.)*<br />

Growth of scientific staff (in%)*<br />

Growth of scientific staff (not employed) (in%)*<br />

Average duration of scientific staff<br />

Expenses <strong>for</strong> training<br />

Structural <strong>Capital</strong><br />

Investments in library and electronic medias*<br />

Relational <strong>Capital</strong><br />

Research grants abroad (as % of scientific staff)*<br />

International scientists at the university (total in months)*<br />

Number of conferences visited*<br />

Number of employees financed by non-institutional funds<br />

Number of activities in committees etc.<br />

Hit rate EC <strong>research</strong> programs<br />

New co-operation partners<br />

Research<br />

Publications (referred)*<br />

Publications (proceedings etc.)<br />

Publications total*<br />

Number publications with co-authors from the industry<br />

Habilitation*<br />

PhDs<br />

Non-institutional funds (contract <strong>research</strong> etc.)<br />

Education<br />

Graduations<br />

Average duration of studies<br />

Teacher per student<br />

Drop-out-ratio<br />

PhDs and master theses finalised<br />

Commercialising<br />

Number of spin-offs<br />

Employees created by spin-offs<br />

Income generated from licences<br />

Knowledge transfer to the public<br />

Hits internet site<br />

Lectures (non-scientific)<br />

Services<br />

Measurement and lab services and expert opinions<br />

Leasing of rooms and equipment<br />

20

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!