27.06.2013 Views

6th European Conference - Academic Conferences

6th European Conference - Academic Conferences

6th European Conference - Academic Conferences

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Proceedings<br />

of the<br />

2nd International<br />

<strong>Conference</strong> on Information<br />

Management and<br />

Evaluation<br />

Ryerson University<br />

Toronto<br />

Canada<br />

27-28 April 2011<br />

Edited by<br />

Professor Kenneth Grant<br />

Ryerson University<br />

Toronto<br />

Canada


Copyright The Authors, 2011. All Rights Reserved.<br />

No reproduction, copy or transmission may be made without written permission from the individual authors.<br />

Papers have been double-blind peer reviewed before final submission to the conference. Initially, paper<br />

abstracts were read and selected by the conference panel for submission as possible papers for the<br />

conference.<br />

Many thanks to the reviewers who helped ensure the quality of the full papers.<br />

These <strong>Conference</strong> Proceedings have been submitted to Thomson ISI for indexing.<br />

Further copies of this book and previous year’s proceedings can be purchased from http://academicconferences.org/2-proceedings.htm<br />

ISBN:97-1-906638-97-9 Book<br />

Published by <strong>Academic</strong> Publishing International Limited<br />

Reading<br />

UK<br />

44-118-972-4148<br />

www.academic-publishing.org


Contents<br />

Paper Title Author(s) Page<br />

No.<br />

Preface v<br />

Biographies of <strong>Conference</strong> Chairs,<br />

Programme Chair, Keynote Speaker and<br />

Mini-track Chairs<br />

Biographies of contributing authors vi<br />

Performance Management of the NHS’<br />

Mental Health Care Service Delivery in<br />

England: The Role of the Service’ Actors’<br />

(Clients) Collaboration Through Dialogism<br />

The Acceptance and Use of Educational<br />

Technology in Kingdom of Bahrain<br />

Comparison of Feature Selection Techniques<br />

Using Fully-Controlled Simulation-Based<br />

Datasets<br />

A Delphi-Multi-Criteria Decision Making<br />

Approach in the Selection of an Enterprise-<br />

Wide Integration Strategy<br />

Selection, Implementation and Post<br />

Production of an ERP System<br />

Evolution of eProcurement Model in the<br />

Italian Public Sector: From Government<br />

Centralisation to Regional Delocalisation<br />

ICTs as Weapons of Mass Interaction -<br />

Motivations and Implications of Mediated<br />

Control<br />

Investigating the Factors Inhibiting SMEs<br />

from Recognizing and Measuring Losses<br />

From Cyber Crime in South Africa<br />

Kwesi Korsa Aggrey 1<br />

Jaflah Al-Ammary 9<br />

Suzan Arslanturk, Mohammad Siadat,<br />

Theophilus Ogunyemi , Ishwar Sethi and<br />

Ananias Diokno<br />

Tridip Bardhan, James Ngeru, and Richard<br />

Pitts<br />

Imran Batada and Asmita Rahman 38<br />

Clara Benevolo and Renata Paola Dameri 45<br />

Kofi Agyenim Boateng 54<br />

Gino Bougaardt and Michael Kyobe 62<br />

Sharing Knowledge – The CoP way Sheryl Buckley and Apostolos Giannakopoulos 72<br />

An Empirical Framework of key Success<br />

Factors for Software Process Improvement<br />

Assessing Information Management<br />

Competencies in Organisations<br />

A Semiotic Approach to Analyse the<br />

Influencing Factors in Knowledge Transfer<br />

Exploring the Emotional Exhaustion of<br />

Healthcare Providers Using an In-Hospital<br />

Employee Communication Network<br />

The Implementation of RSS-Based Clinical<br />

Reminders<br />

Factors Determining the Adoption of ICTs in<br />

Healthcare Service Delivery: A Developing<br />

Country Context<br />

Evaluation of Application Embedded<br />

Knowledge Migration Issues<br />

Nuntarat Bunditwongrat, Mathupayas<br />

Thongmak and Atcharawan Ngarmyarn<br />

Andy Bytheway 91<br />

Supaporn Chai-Arayalert and Keiichi Nakata 103<br />

Cheng-Yi Chiang, Ya-Ying Kuo, Ying-Hui Hou<br />

and I-Chiu Chang<br />

Wen-Chou Ch, Chia-Hsien Wen, Sek-Kwong<br />

Poon and Shih-Che Huang 3<br />

Taurai Chikotie, Jonathan Oni and Vesper<br />

Owei<br />

vi<br />

18<br />

24<br />

82<br />

112<br />

120<br />

127<br />

Mitchell Cochran 134<br />

i


Paper Title Author(s) Page<br />

No.<br />

Motivations for IT Outsourcing in Public<br />

Sector Local Government<br />

Documenting Innovation: A Methodological<br />

Proposal and Application<br />

Legal, Privacy, Security, Access and<br />

Regulatory Issues in Cloud Computing<br />

Interoperability Monitoring for eGovernment<br />

Service Delivery Based on Enterprise<br />

Architecture<br />

Enterprise Resource Planning<br />

Implementation Differences Within the Same<br />

Methodology - Case Study From West<br />

Europe and Turkey<br />

A Meta-Analysis of how the Adoption of<br />

Innovative Web 2.0 Tools Like Digital<br />

Business Ecosystems can Lead to Improved<br />

SME Collaboration<br />

Ideas About Profitability in Research and<br />

Development and the Selective Pressure<br />

From Management Accounting<br />

Case Study on Information Evaluation by GIS<br />

for Aging Society Urban Planning: GIS<br />

Application on Urban Planning-<br />

Distortion Free Algorithm to Handle<br />

Secondary Watermark Attack in Relational<br />

Databases<br />

Engineering Change Through the Domains of<br />

Enterprise Architecture<br />

Evaluating Interoperability Artifacts for the<br />

Exchange of Public Service Information:<br />

Outline of a Conceptual Framework<br />

Free and Open Source Software for Public<br />

Sector Enterprise Applications in Sri Lanka<br />

Evaluating the Success of Customer<br />

Relationship Management (CRM) Systems<br />

Toward a Novel Methodology for IT Strategic<br />

Planning<br />

Applying an Innovative Information System to<br />

Improve the Effects of Community Health<br />

Promotion<br />

Information Security Practices in Zambian<br />

Copper Mines: An Investigation Into the<br />

State-of-Practice of Information Security<br />

Within Zambian Copper Mines Based on the<br />

ISO/IEC 27002 Standard<br />

Proposing an Optimized Change<br />

Management Process by Analyzing ITSM<br />

Frameworks<br />

Michael Cox, Martyn Roberts and John Walton 141<br />

Stefano De Falco 150<br />

Nomusa Dlodlo 161<br />

Badr Elmir, Nabil Alrajeh and Bouchaib<br />

Bounabat<br />

169<br />

Turan Erman Erkan 181<br />

Francesc Estanyol 187<br />

Albrecht Fritzsche 196<br />

Hiroatsu Fukuda, Yupeng Wang and Kiyoshi<br />

Shinriki<br />

Sajid Iqbal, Azhar Rauf, Huma Javed and<br />

Shabir Ahmad<br />

204<br />

214<br />

Tiko Iyamu 222<br />

Veit Jahns 231<br />

Srimal Jayawardena and Gihan Dias 239<br />

Farnoosh Khodakarami and Yolande Chan 253<br />

Mouhsine Lakhdissi and Bouchaib Bounabat 263<br />

Chih-Yu Lin, Wen-Yu Ho, Chia-Hsien Wen and<br />

Hui-Mei Hsu<br />

274<br />

Chishala Lukwesa and Christopher Upfold 281<br />

Negar Madani, Ali Suzangar, Mohamad Kajbaf,<br />

Shirin Nasher and Mehrdad Kalantarian<br />

ii<br />

291


Paper Title Author(s) Page<br />

No.<br />

The use of RFID and Web 2.0 Technologies<br />

to Improve Inventory Management in South<br />

African Enterprises<br />

Modeling the Genetic Schemes of Human<br />

Memory Information in the Process of<br />

Production and Reparation of Knowledge<br />

Copyright Protection for GIS Vector Map<br />

Based on Wavelet Transform<br />

Evaluation of IT Investment Methods and<br />

Proposing a Decision Making Model<br />

Critical Factors in the use of Mobile Learning<br />

by “Digital Natives” on Portuguese Teaching<br />

A Framework for Information Systems<br />

Innovation: A Case of Competitive<br />

Intelligence in Organisations<br />

RFID: A Review of its Relevance and<br />

Application in South African Retailing<br />

Systems<br />

Towards a Comprehensive Evaluation<br />

Framework for ICT for Development<br />

Evaluation – an Analysis of Evaluation<br />

Frameworks<br />

User Adoption of the New Health Information<br />

System - Individual and Organizational<br />

Determinants<br />

Relationship Between Organizational Factors<br />

and RFID Adoption in Iranian Gas Company<br />

Central Warehouses<br />

Business Intelligence as Decision Support in<br />

Business Processes: An Empirical<br />

Investigation<br />

Assessing Future Value of Investments in<br />

Security-Related IT Governance Control<br />

Objectives – Surveying IT Professionals<br />

Evaluation Information Extraction for Health<br />

Text Categories Using C4.5 and Naïve Bayes<br />

Optimizing Information Technology Value<br />

Governance Framework Based on Val IT<br />

Developing an Outpatient Electronic Medical<br />

Record System in Taiwan From a Physicians’<br />

Perspective<br />

MTN Foundation's Digital Library Project in<br />

Nigerian Universities: An Evaluative Study<br />

Cloud Computing-Based IT Solutions for<br />

Organizations With Multiregional Branch<br />

Offices<br />

Case Study on Information Evaluation by GIS<br />

for Aging Society Urban Planning:<br />

Information Evaluation of Shrinking Cities<br />

Sizakele Mathaba, Nomusa Dlodlo, Quentin<br />

Williams and Mathew Adigun<br />

Fatemeh Mohammadi, Tahereh Mohammadi<br />

and Bahram Rezaie<br />

300<br />

308<br />

Amer Sedeeq Mustafa 317<br />

Shirin Nasher, Mehrdad Kalantarian, Ahmad<br />

Akbari, Ali Suzangar, Mohammad Kajbaf and<br />

Negar Madani<br />

324<br />

Mário Carrilho Negas and Paulo Ramos 333<br />

Phathutshedzo Nemutanzhela and Tiko Iyamu 341<br />

Jonathan Oni, Edward Dakora and Vesper<br />

Owei<br />

351<br />

Caroline Pade-Khene and Dave Sewry 360<br />

Bahlol Rahimi 370<br />

Farnaz Rahimi and Gholamabas Arabshahi 378<br />

Ari Riabacke, Aron Larsson and Mats<br />

Danielson<br />

Waldo Rocha Flores, Teodor Sommestad,<br />

Hannes Holm and Mathias Ekstedt<br />

Klaokanlaya Silachan and Panjai<br />

Tantasanawong<br />

Ali Suzangar, Mehrdad Kalantarian , Shirin<br />

Nasher , Mohammad Kajbaf and Negar Madani<br />

Hsiao-Ting Tseng, Pei-Ting Chang and Ray<br />

Chang<br />

384<br />

393<br />

403<br />

412<br />

420<br />

Ngozi Blessing Ukachi 428<br />

Harris Wang 435<br />

Yupeng Wang, Hiroatsu Fukuda and Kiyoshi<br />

Shinriki<br />

iii<br />

441


Paper Title Author(s) Page<br />

No.<br />

The Impact of Software Test Constraints on<br />

Software Test Effectiveness<br />

Strategic Considerations for the Effective<br />

Mapping of Educational Technology to<br />

<strong>Academic</strong> Outcomes<br />

Grafton Whyte and Donovan Lindsay Mulder 450<br />

Hossein Zadeh, Arthur Shelley and Rod<br />

McCrohan<br />

PhD Research Papers 471<br />

Evaluating the Performance of ERP Systems<br />

in Saudi Arabian Higher Education: A<br />

Stakeholders’ Perspective<br />

The Benefits of ICT Adoption: An Empirical<br />

Study of Nigerian SMEs<br />

Establishing the Suitability of Dynamic Web<br />

Applications Development Platforms for<br />

Learning web Techniques in Tertiary<br />

Institutions<br />

Human Resources Transformation Beyond<br />

Boundaries in Outsourcing Business Model<br />

s- Expatriate Benchmarking<br />

The Diffusion of Mobile Phones for Business<br />

and Information Management in Kenya<br />

Querying DTI Analysis Results Within Deep<br />

Perisylvian Area<br />

Towards an Evaluation Method for<br />

Information Quality Management of Health<br />

Information Systems<br />

Developing a Change Management Model for<br />

Iranian <strong>Academic</strong> Libraries: A Delphi Study<br />

Mona Althonayan and Anastasia<br />

Papazafeiropoulou<br />

461<br />

473<br />

Idisemi Apulu and Ann Latham 483<br />

Johnson Dehinbo 492<br />

Swathi Duppada and Rama Chandra Aryasri 502<br />

Wakari Gikenye 511<br />

Sarmad Istephan and Mohammad Siadat 521<br />

Siti Asma Mohammed and Maryati Mohd Yusof 529<br />

Maryam Nakhoda and Sirous Alidousti 539<br />

Adding Action to the Information Audit Huan Vo-Tran 546<br />

Business Intelligence Best Practices for<br />

Success<br />

Joseph Woodside 556<br />

Work in Progress 563<br />

The Evolution of IT Governance in Multiple<br />

Industry Units and the Business Case for an<br />

Outside Partner<br />

The Relationship Between Quality<br />

Management and Knowledge Management in<br />

the Service Industries<br />

Construction and Validation of eSchool<br />

Success Model<br />

Ganeshprasad Chandrasekaran,<br />

Chandramohan Annavarapu and Balaji<br />

Rathinasamy<br />

565<br />

Amir Honarpour and Ahmad Jusoh 568<br />

Hesbon Nyagowa, Dennis Ocholla and<br />

Stephen Mutula<br />

iv<br />

572


Preface<br />

Following on from the continued success of the <strong>European</strong> <strong>Conference</strong> on Information Management and<br />

Evaluation, we are delighted at the Ted Rogers School of Management, Ryerson University to be able to<br />

host the 2nd International <strong>Conference</strong> on Information Management and Evaluation (ICIME 2011).<br />

ICIME aims to bring together individuals researching and working in the broad field of information<br />

management, including information technology evaluation. We hope that this year’s conference will provide<br />

you with plenty of opportunities to share your expertise with colleagues from around the world.<br />

This year’s opening keynote address will be delivered by Dr Catherine Middleton, Ted Rogers School of<br />

Information Technology Management, Ryerson University, Toronto, Canada.<br />

A second keynote address on day two will be delivered by Gary Davenport, Vice-President of Information<br />

Technology, AllStream Inc, Canada.<br />

ICIME 2011 received an initial submission of 180 abstracts. Of this 65 papers have been accepted for these<br />

<strong>Conference</strong> Proceedings as a result of the double blind, peer review process of full paper submissions.<br />

These papers come from all parts of the world including Australia, Canada, India, Iran, Italy, Japan,<br />

Malaysia, Morocco, Nigeria, Pakistan, Portugal, South Africa, Sri Lanka, Sweden, Thailand, Taiwan, the<br />

United Kingdom, the United States of America Zambia and Zimbabwe. So ICIME is set to be a truly<br />

international conference!<br />

We wish you a most interesting and enjoyable time in Canada<br />

Ken Grant<br />

<strong>Conference</strong> Chair<br />

April 2011<br />

v


Biographies of <strong>Conference</strong> Chairs, Programme Chairs<br />

and Keynote Speakers<br />

<strong>Conference</strong> Chair<br />

Ken Grant is an associate professor in the Ted Rogers School of Management at<br />

Ryerson University, His teaching and research interests are in the areas of Business<br />

& Technology Strategy, Innovation & Knowledge Management and Electronic<br />

Commerce. He holds the Ted Rogers School of Management Faculty Teaching Chair<br />

and is a former Director of the School of Information Technology Management at<br />

Ryerson. Prior to becoming part of the Ryerson community, Ken spent over 30 years<br />

in industry, mainly as a management consultant. He has held senior partner/vice<br />

president positions in major consulting firms, including A.T. Kearney, EDS and KPMG.<br />

He holds a BA from The Open University, UK., an MBA from York University, Toronto<br />

and is just completing his DBA at Henley Business School, University of Reading.<br />

Keynote Speakers<br />

Dr Catherine Middleton holds a Canada Research Chair in Communication<br />

Technologies in the Information Society. Her research focuses on the<br />

development and use of new communication technologies, with specific interests<br />

in mobile devices and fixed and wireless broadband networks. She is also<br />

interested in how Canadians use (or don't use) the internet in their daily lives. A<br />

member of the GRAND Networks of Centres of Excellence projece, she is the<br />

Leader of the New Media Challenges and Opportunities research theme and also<br />

leads the GRAND Digital Infrastructures project. Catherine teaches Innovation and<br />

Organization Theory in the MBA/MMSc program in the Ted Rogers School of<br />

Management, and teaches and supervises students in the Communication and<br />

Culture graduate program.<br />

Gary Davenport As Vice-President of Information Technology, Gary is responsible for<br />

the definition and implementation of transformational business initiatives using<br />

information technology for the Allstream business unit of MTS Allstream Inc. In this<br />

role, Gary leads a team of dedicated IT professionals focused on improving business<br />

capabilities in a very competitive telecommunications industry. He is also the<br />

Executive Sponsor for Allstream's Workplace 2.0 Program enabling much greater<br />

levels of employee mobility, productivity and engagement. Gary has participated on<br />

many North American and <strong>European</strong> CIO Councils and also served on the Board of<br />

Directors of the Voluntary Inter-industry Communications Standards (VICS)<br />

organization. He is currently a Board<br />

Member of the CIO Association of Canada. Gary holds a Bachelor of<br />

Arts<br />

degree from York University.<br />

Biographies of contributing authors (in alphabetical order)<br />

Kwesikorsa Aggrey is a PhD student of the University of Lincoln, UK. Kwesikorsa research area is in health<br />

service delivery, mental health service delivery in particular. His research topic is on collaboration between<br />

the NHS Trusts (Lincolnshire Partnership Foundation<br />

Trusts) and the Voluntary Sector in Performance of<br />

Mental<br />

Health Care Delivery in Lincolnshire.<br />

Jaflah Hassan AlAmmary is assistant professor in the Information System Department of the college of the<br />

IT at University of Bahrain. She holds PhD from University of Murdoch. Al-Ammary’s research interest<br />

focuses on Strategic alignment, Knowledge Management, and E-learning and Educational<br />

Technology. She<br />

has<br />

published many papers on these topics on conferences, journals and books.<br />

Mona Althonayan is a research student at the school of information<br />

systems, computing and mathematics,<br />

Brunel<br />

University, UK, working on ERP systems evaluation.<br />

Al-Mallah Amer is an Assistant Professor at the College of Science, he obtained a Ph.D. degree in<br />

Computer Science. Then he worked at the University of Technology. Subsequently, served as M.Sc.<br />

students supervisor at the postgraduate studies department. Dr. Amer<br />

has published eleven journal articles.<br />

His<br />

publications reflect his research interests in information security.<br />

vi


Idisemi Apulu is currently a PhD student at the School of Technology, University of Wolverhampton, United<br />

Kingdom where she also obtained her Masters degree in Strategic Information Technology Management.<br />

She also holds a BSc degree in Computer Science from the University of Abuja, Nigeria.<br />

Suzan Arslanturk is a masters student at the Department of Computer Science, Oakland University. She<br />

holds a B.S in Computer Engineering. Her research interests are image processing, pattern recognition and<br />

data mining. Suzan‘s master research focused on data mining and machine learning.<br />

Rahimi Bahlol, PhD: studied informatics at Linköping University. He is now assistant Professor at Urmia<br />

University of Medical Sciences. His research interests are; Individual, organizational and social issues and IT<br />

solution, collaborative work through ICT, IT solutions in health care setting, implementation and use and<br />

evaluation of integrated HIS, Adoption of IT solution.<br />

Tridip Kumar Bardhan is working as Chairperson and Director of Manufacturing Process Laboratory at the<br />

Department of Industrial and Systems Engineering (ISE) of Morgan State University in Baltimore, Maryland,<br />

USA. He received his BSc degree in Industrial Arts from Dhaka University and BSIE, MSIE and PhD in<br />

Industrial Engineering from Wichita State University, USA.<br />

Imran Batada is a senior IT Professional with over 10 years of experience in the IT industry. As an IT<br />

Professional he has worked in USA and Asia. He is currently working in Institute of Business Administration,<br />

Karachi Pakistan as Head of Information System Department. He is also responsible for overall ERP<br />

Implementation.<br />

Kofi Agyenim Boateng completed his PhD in Information Systems at the London School of Economics and<br />

Political Science, UK, in December 2009 and, currently, a lecturer at the British Institute of Technology & Ecommerce.<br />

Kofi’s research interests fundamentally lie in the innovative application of ICTs and their<br />

psychological consequences in organisational strategising.<br />

Sheryl Buckley is Deputy Head of Department in the Department of Business Information Technology (BIT)<br />

at the University of Johannesburg (UJ). Her passion lies in the Information Science discipline. She is a<br />

committee member of a number of international organisations as well as an active peer reviewer. She has<br />

presented and published papers locally and internationally.<br />

Nuntarat Bunditwongrat received her Bachelor’s Degree (2005) in Information Technology for Business<br />

from Chulalongkorn University, Thailand. She is now a Master’s degree student of Management Information<br />

Systems program at the Faculty of Commerce and Accountancy, Thammasat University. She also serves as<br />

a Senior Consultant at the Department of Global Business Services, IBM Thailand Company Limited.<br />

Supaporn Chai-Arayalert is a Ph.D. student in Informatics at Informatics Research Centre, Henley<br />

Business School, University of Reading, UK. Her research areas focus on knowledge management,<br />

semiotic approach and Green ICT. She graduated Master of Science in Management of Information<br />

Technology and she recently works as lecturer at Prince of Songkhla University, Thailand<br />

Yolande Chan is a Professor, MIS at Queen’s School of Business and Director, The Monieson Centre. She<br />

holds a Ph.D. from the Richard Ivey School of Business, an M.Phil. in Management Studies from Oxford, and<br />

S.M. and S.B. degrees in Electrical Engineering and Computer Science from M.I.T. She is a Rhodes Scholar.<br />

Dr. Chan conducts research on knowledge management and information technology strategy.<br />

Ganeshprasad Chandrasekaran B.E., M. Tech., M.B.A., is an associate of India’s largest Software<br />

Company TCS Limited, where he is focused with aligning the IT goals with Business priorities. He has been<br />

awarded with Chief Minister Award and University Gold medal for the best academic excellence. Ganesh<br />

has been a research scholar in SRM University, India.<br />

Taurai Chikotie is a student and researcher at the Cape Peninsula University of Technology in the fields of<br />

ICT and Healthcare for Development. He holds a BSc Honors degree in Information Science from the<br />

National University of Science and Technology in Zimbabwe and a Masters Degree in Information<br />

Technology cum laude from the Cape Peninsula University of Technology.<br />

Wen-Chou Chi is a Doctoral Student in the MIS program at the National Chung Cheng University in Taiwan.<br />

His current research interests include hospital information systems, knowledge management and decision<br />

support system.<br />

vii


Mitchell Cochran has been the Information Systems Manager for the City of Monrovia for 14 years. He has<br />

also worked in the Court System and for IBM. He is currently working on his Information Systems PhD from<br />

Claremont Graduate University and has completed Masters Degrees in Administration and Homeland<br />

Security. He has a CISM certification.<br />

Renata Paola Dameri is senior lecturer in Business Administration at the Faculty of Economics, University of<br />

Genova, Italy. She is professor in Accounting and Information Systems in Genova, visiting professor in IT<br />

Governance at Université de Paris Dauphine and Fellow of the SDA Bocconi School of Management, Milano,<br />

Italy. Her research interest covers IT investments evaluation, IT governance, IT security and compliance.<br />

Mats Danielson is a full Professor in Computer and Systems Sciences and Vice Dean of the Social Science<br />

Faculty at Stockholm University. He holds a Ph.D in Computer and Systems Sciences as well as degrees in<br />

Business Administration and Computer Engineering. He has been working with decision and risk analysis<br />

professionally and academically for almost 20 years.<br />

Stefano De Falco Holds a Degree and a Ph.D degree in electrical engineering, University of Naples<br />

Federico II. His research interests concern the modelling and testing of the measures for quality of firms. His<br />

publications are present in international congresses and journals. He is the inventor and the advisor of a<br />

national magazine “TT-Techology Transfer”.<br />

Johnson Dehinbo obtained B.Sc. degree in Computer Science & Statistics from Ogun State University in<br />

1989, and B.Sc. Honors in Information Systems from UNISA in 2000. He obtained M.Sc. degree in<br />

Information Systems specializing in software engineering from UNISA in 2006. He is currently busy his<br />

M.Phil/Ph.D studies in Informatics at the University of Pretoria.<br />

Nomusa Dlodlo holds a PhD in Computer Science. She is currently employed as a Senior Researcher at<br />

the Council for Scientific and Industrial Research (CSIR)'s Meraka Institute in Pretoria, South Africa. Dr.<br />

Dlodlo works in the Internet-of-Things (IoT) group. Her main research interests lie in applying IoT<br />

technologies to improve teaching and learning at primary and secondary school levels<br />

Swathi Duppada received her Master’s degree in Human Resource Management from Andhra University,<br />

Visakhapatnam, India, and is currently working with Satyam Computer Services Limited, Canada. She is<br />

currently a Ph.D. candidate at Jawaharlal Nehru Technological University, Hyderabad, India. Her research<br />

interests include Benchmarking of Human Resource Management, Modernization of HR using ERP and<br />

Software Testing.<br />

Badr Elmir is a Software Engineer graduated from ENSIAS (2002), holder of an Extended Higher Studies<br />

Diploma from ENSIAS (2006) and is a “Ph.D. candidate” at ENSIAS since 2009. His research focuses on<br />

interoperability monitoring within public administration. He is an integration architect on the Ministry of<br />

Economy and Finance of Morocco since 2002.<br />

Turan Erman Erkan is an Assistant Professor at the Department of Industrial Engineering in Atılım<br />

University. After a career in management consultancy, he worked as an international SAP consultant and<br />

then started a research career. His research and consultancy interests include ERP, BPR, SCM, CRM and<br />

performance measurement.<br />

Francesc Estanyol is research assistant at the University of Edinburgh Business School. He received his<br />

degree in Computer Engineering from the Universitat Pompeu Fabra, Spain, and also holds masters degrees<br />

in Bioinformatics for Health Sciences and in Management. Currently he is the Scientific Manager of<br />

EcoBusiness Marie Curie project. His research interests include how the adoption of innovative ICT tools can<br />

benefit SMEs and the translation of research into real market applications.<br />

Albrecht Fritzsche has a Master's degree in mathematics, and educational science, D.Phil. Vehicle<br />

scheduling expert in the IT Management department of Daimer, A.G., Stuttgart. Doctoral dissertation in<br />

industrial economics on heuristic search in complex environments, doctoral dissertation in philosophy on<br />

indeterminacy in technical systems. Interdisciplinary research on innovation at TU Darmstadt.<br />

Hiroatsu Fukuda is a professor at the University of Kitakyushu of Japan. A doctor of engineering, registered<br />

architect of Japan, member of Architectural Institute of Japan (AIJ). Research interests are environment<br />

design of architecture, recyclable buildings, and urban planning for compact cities, urban environmental<br />

evaluation and suggestion.<br />

viii


Waldo Rocha Flores received his B.Sc. in Business Administration and Economics from the Stockholm<br />

University in 2007 and his M.Sc. in Electrical Engineering from the Royal Institute of Technology (KTH) in<br />

Stockholm in 2008. He is now pursuing a Ph.D. in Industrial Information and Control Systems.<br />

Wakari Gikenye, B.A. M.A. University of Nairobi, PGD University of Wales, is a Senior Librarian at the<br />

University of Nairobi Library, Nairobi, Kenya, and a PhD Student at University of Zululand, South Africa. Has<br />

recently presented papers on the diffusion of Information and Communication Technologies in the informal<br />

sector in Kenya.<br />

Amir Honarpour is currently a PhD. Student at Faculty of Management and Human Resource Development<br />

(FPPMS) University Technology Malaysia. Amir graduated with a M.Sc. in system management in 2009.<br />

After graduation, he worked in a project to help Design an integrated research system among Iran's<br />

universities. His research interests include: Knowledge Management, Web-based Courseware Tools, Quality<br />

Management and Research information Systems.<br />

Sajid Iqbal received his Masters degree in computer science from University of Peshawar Pakistan.<br />

Currently he is a research student at Department of computer science, university of Peshawar Pakistan. His<br />

area of interest includes information security, relational database watermarking and data mining.<br />

Sarmad Istephan is a Computer Science PhD student at Oakland University. His research focuses on the<br />

storage and querying of Medical Images (e.g., MRI). In addition to being a PhD student, Istephan was a<br />

Java/C# Software Engineer and currently is a Senior SQL Server Database Engineer at Quicken Loans.<br />

Istephan eagerly awaits publishing more scientific papers in his field.<br />

Tiko Iyamu is a Professor of Information Systems at the Tshwane University of Technology, Pretoria. His<br />

research interests include Mobile Computing, Enterprise Architecture and Information Technology Strategy.<br />

Theoretically, he focuses on Actor Network Theory (ANT) and Structuration Theory (ST). Iyamu is author of<br />

numerous peer-reviewed journal and conference proceedings articles.<br />

Veit Jahns has a German Diploma in Computer Science and works as a software developer and consultant<br />

at the otris software AG in Dortmund, Germany. Additionally, he is finishing his Ph.D. thesis at the University<br />

of Duisburg-Essen. His research interests are the all aspects of information systems interoperability, in<br />

particular between information systems in public authorities.<br />

Srimal Jayawardena obtained his BSc Engineering from the University of Peradeniya and BIT from the<br />

University of Colombo School of Computing, both with first class honours. He has served in the Central Bank<br />

of Sri Lanka as an Assistant Director/IT d and at the Information and Communication Technology Agency of<br />

Sri Lanka as a Technology Specialist. He is currently a PhD candidate at the Australian National University.<br />

Mehrdad Kalantarian is ITSM specialist of Infoamn CO., a consulting firm that provides services and<br />

solutions for security, compliance & IT Management. He works with a professional team studying on IT fields<br />

such as ITIL, ISMS, COBIT and Val IT. His experimental field is COBIT. He has M.S. degree in ICT<br />

engineering and lives in Tehran.<br />

Farnoosh Khodakarami has an MSc in Management from Queen’s School of Business, Queen's University,<br />

Canada. Currently, she is working as a researcher at The Monieson Centre, Queen’s University. Her<br />

research interests include customer relationship management, e-commerce, information systems and<br />

knowledge management.<br />

Ya-Ying Kuo is a master of Healthcare Information Management at National Chung Cheng University,<br />

Chiayi, Taiwan. Her current research interests include hospital information systems and patient privacy.<br />

Michael Kyobe is an associate Professor in the department of Information Systems, University of Cape<br />

Town. He holds a PhD and an MBA. Prior to joining academia, Michael worked for over 15 years in the<br />

public and private sectors in the IS and Computer forensics fields. His research include information security,<br />

business alignment and KM<br />

Mouhsine Lakhdissi is a Partner in a consulting firm specialized in IT Architecture. He received a Master<br />

degree in Software Engineering in a leading engineering school in Morocco. He also served as Chief IT<br />

Architecture in many companies with international exposure. His research interests include Enterprise<br />

Architecture, IT Governance and processes and Software industrialization.<br />

ix


Aron Larsson, Ph.D. in Computer and Systems Science and MBA. Senior lecturer and researcher at<br />

Stockholm University and Mid Sweden University. Research interests include methods, procedures and<br />

applications of computer based decision support, as well as risk and decision analysis. Research projects<br />

include process models and methods for public decision making, landmine clearance activities, procurement<br />

processes, and distributed artificial intelligence in wireless networks.<br />

Chih-Yu Lin is a Doctoral Student in the MIS program at the National Chung Cheng University in Taiwan.<br />

His current research interests include hospital information systems, knowledge management and decision<br />

support system<br />

Sizakele Untonette Mathaba has competed her degree in Bsc Information Systems (2007) honours degree<br />

in Computer Science (2008) at the University of Zululand. In 2009, she joined CSIR (Council for Scientific<br />

and Industrial Research) as an internship student. Currently, she is registered for her Masters degree in<br />

Computer Science. She is working with the Internet of Things Engineering Group at CSIR (Meraka).<br />

Fatemeh Mohammadi has a Ph.D in Educational management. Faculty member of Islamic Azad University-<br />

Shiraz Branch, Iran.He is an inventor with 5 registered inventories. The author of 6 books and 24 essays<br />

articles and papers with 12 presentations in National <strong>Conference</strong>s. Instructor of 38 educational terms for<br />

faculty members and Theory builder of Human Information Life.<br />

Siti Asma Mohammed is currently a PhD student at National University of Malaysia. She finished Masters of<br />

IT specializing in Information Systems at University of Sydney, Australia. She worked as a Test Engineer for<br />

one year and Assistant Lecturer in Information Systems for four years before pursuing PhD. Her research<br />

interests are IS Evaluation and information quality.<br />

Maryam Nakhoda is a PhD candidate in Library and Information Science (LIS) at University of Tehran,<br />

faculty of Psychology and Education. She is the author of papers in Persian and English. Her research<br />

interests include Information Technology (IT) application in academic libraries, library management, and<br />

managing change in academic libraries.<br />

Shirin Nasher is ITSM specialist of Infoamn CO., a consulting firm that provides services and solutions for<br />

security, compliance & IT Management. She works with a professional team studying on IT fields such as<br />

ITIL, ISMS and COBIT. Her experimental field is Val IT. She has M.S. degree in IT management and lives in<br />

Tehran.<br />

Mário Carrilho Negas is an assistant professor of management at the Open University (Portugal). He<br />

received his Ph.D. in Management. His main research interests include the adoption of systems and<br />

information technology in SMEs, strategic planning of information systems and management of Innovation.<br />

Phathutshedzo Nemutanzhela is a Masters student at the Tshwane University of Technology. She has a<br />

Baccalaureus Technologies (BTech): Information Technology (Informatics). Her principle research interest is<br />

Competitive Intelligence and Information Systems.<br />

Hesbon Nyagowa is a Ph.D student at the University of Zululand, South Africa specializing in Information<br />

Studies. He obtained Master of Business and Administration specializing in Management Information<br />

Systems at University of Nairobi Kenya in 2002. His Bachelor’s degree was in Education (Science) obtained<br />

at Kenyatta University, Kenya in 1988. Is currently the <strong>Academic</strong> Registrar, Kenya Polytechnic University<br />

College.<br />

Jonathan Oni is a Post Graduate student and researcher at the Cape Peninsula University of Technology,<br />

Cape Town, South Africa, with an interest in e-business. He holds a BSc Honors in Computer Science.<br />

Jonathan consults for various Information Technology companies and involved in managing IT projects. He<br />

is also a part-time lecturer at a higher educational institution.<br />

Farnaz Rahimi is studying IT management in Alzahra University(MS degree) and work in contract<br />

department in Mashhad Gas Company. Farnaz is interested in the field of "knowledge management" and<br />

"implementing new IT technologies in organization"<br />

Azhar Rauf received his doctorate degree in computer science from Colorado Technical University,<br />

Colorado Springs CO, USA in 2007. Currently he is teaching as Assistant Professor at the Department of<br />

Computer Science, University of Peshawar, Peshawar, Khyber Pakhtunkhwa, Pakistan. His areas of interest<br />

include Relational Database Watermarking, Fine-Grained Security techniques in Relational Databases,<br />

Encryption, Anonymity and Information Security.<br />

x


Ari Riabacke, Ph.D. in Risk- and Decision Analysis (Computer Science), Head of Business Intelligence at<br />

the largest Swedish Management and IT Consultant Company, also has a M.Sc. in Organizational Decision<br />

Making, and is a member of the DECIDE Research Group at Stockholm University.<br />

Martyn Roberts spent the first few years of his career in industry working in information systems, but<br />

transferred to academia over 20 years ago. He is now Principal Lecturer at the University of Portsmouth. He<br />

has taught various aspects of IS on a wide of programmes at both undergraduate and post graduate levels.<br />

He has published mainly in the areas of strategic information systems and eCommerce.<br />

Klaokanlaya Silachan works at the Computer technology department, Facultly of Science Nakorn pathom<br />

Rajabhat University, Thailand. She is currently pursuing in computer information technology, Silpakorn<br />

University,Thailand major in data mining, Health medical data, ontology. She received her master’s degree in<br />

Information Management Technology from Mahidol University, Thailand, in 1998. She has publication in<br />

national conference and international conference proceedings.<br />

Ali Suzangar is a Chief System Officer of Infoamn CO., a consulting firm that provides services and<br />

solutions for security, compliance & IT Management. He works with a professional team studying on IT fields<br />

such as ITIL, ISMS and COBIT. His experimental field is Val IT. He has M.S. degree in IT management and<br />

lives in Tehran.<br />

Panjai Tantassanawong is currently pursuing his doctoral degree in Computer Science majoring in<br />

networking and software engineering at AIT . He received his master’s degree in Computer science from<br />

Chulalongkorn University, Thailand in 1992. He is Assistant Professional the Computing Program, Facultly<br />

of Science, Silpakorn University, Thailand. He has publications in national and international conference<br />

proceedings<br />

Hsiao-Ting Tseng is a graduate student of Healthcare Information Management at National Chung Cheng<br />

University, Chiayi, Taiwan. Her current research interests include patient privacy, electronic medical records<br />

and exchange of medical images.<br />

Ngozi Ukachi is a librarian at University of Lagos, and presently doing her PhD at University of Nigeria,<br />

Nsukka. A member of International Federation of Library Associations (IFLA), American Library Association<br />

(ALA) and, Nigerian Library Association (NLA). She was IFLA 2010 Essay Competition Award Winner<br />

(organized by IFLA <strong>Academic</strong> and Research Section).<br />

Chris Upfold is currently a lecturer in the Department of Information Systems at Rhodes University, South<br />

Africa. He also teaches in the Rhodes Business School. His areas of interest and research are Information<br />

Security, Radio Frequency Identification (RFID), Project Management, Virtual Teams and Corporate<br />

Communications.<br />

Huan Vo-Tran is a lecturer and the program director of the Bachelor of Business (Information and<br />

Knowledge Management) within the School of Business IT & Logistics at RMIT University. He is currently<br />

completing a PhD in business computing. Prior to becoming an academic he worked in various fields, which<br />

included project management, systems analysis and high school teaching. His areas of interest include<br />

information management and Web 2.0.<br />

Harris Wang is an associate professor in the School of Computing and Information Systems at Athabasca<br />

University, Canada. He received a PhD in computer science from the Australian National University,<br />

Australia. His research interests include advanced technology for education, information systems and<br />

information security.<br />

Joseph Woodside is a Doctoral candidate in Information Systems at Cleveland State University, with<br />

publications and research interests in topics of business intelligence, informatics, healthcare systems<br />

integration, geo-spatial-temporal modeling, HIT adoption, machine learning, and object-oriented database<br />

technology. Joseph is employed with KePRO, a national care management company, as the Director of<br />

Healthcare Informatics and Business Intelligence.<br />

Hossein Zadeh has taught undergraduate and postgraduate courses in Australia, Hong Kong, Vietnam,<br />

Singapore, and Sweden. Hossein is the recipient of 2008 University Team Teaching Award and 2009<br />

University Certificate of Achievement in innovative teaching. In 2004, Hossein was a visiting scholar at<br />

Linkoping University, Sweden, and in 2009/2010, was a Distinguished Visiting Scholar at IBM Almaden<br />

Research Labs, Silicon Valley (San Jose), USA. Hossein is the recipient of the prestigious 2010 IBM Faculty<br />

Award.<br />

xi


Performance Management of the NHS’ Mental Health Care<br />

Service Delivery in England: The Role of the Service’ Actors’<br />

(Clients) Collaboration Through Dialogism<br />

Kwesi Korsa Aggrey<br />

University of Lincoln, UK<br />

kaggrey@lincoln.ac.uk<br />

Abstract: A welfare state has a responsibility to provide health and social services to the citizenry. The state<br />

delegates that responsibility to a number of actors, each of which has its own performance management criteria. To<br />

ensure coherence, it is important to managing their performances collectively taking into consideration all the actors<br />

inputs rather than only that of government or its agent, the NHS This turns out to be difficult task especially that of the<br />

mental health care services. The author will be looking at a possible solution to this problem based on the<br />

assumption that it is possible to change the way the actors present their stories by bringing them together to share<br />

their purposes and common action through the notion of dialogism as way of maintaining plurality of logics: different<br />

voices (polyphonic), styles (stylistic), space-time conception (chronotopic), interanimating discourse (architectonics)<br />

and the dynamic interplay of different dialogisms (polypi). It is this process of plurality of logic that the researcher<br />

termed, the Third Cybernetics Evolution, as a way of sequential processes from the First Cybernetics through the<br />

Second Cybernetics to the Third Cybernetics. It is argued that implementing these, allows for improved<br />

communication among actors, as a way of achieving high quality service. It is suggested that implementation of these<br />

concepts and processes implies the use of storytelling as a facilitatory system. It is shown that all actors in the mental<br />

healthcare services delivery make use of this system, albeit often inefficiently. Therefore may lead to dissipation of<br />

the system in the future. To prevent such dissipation, the existing structure needs to be improved through spiral<br />

relationships via communication (dialogism)<br />

Keywords: third cybernetics evolution, mental health care system, NHS, collaboration and dialogism<br />

1. Introduction<br />

In qualitative research, one of the issues is how to explore the meaning that actors bring and generate<br />

within social interactions. However, in most researches on mental health care services, the concentration<br />

has been on the inputs of the strong actors namely the government, the NHS and the voluntary<br />

organisations with less attention to the weaker actors like service users, the family, or relatives, carers<br />

and advocates. This paper will focus not only on moving the ‘gear’ to higher level of service users’<br />

controlled and ‘emancipatory’ approach to research but further to the level of dialogism where all<br />

stakeholders are involved in a collaborative effort.<br />

The state through its agent, the NHS, is responsible for ensuring that the health needs of the citizenry are<br />

adequately met. The NHS sometimes sub-contracts some of the services to other organisations (actors,<br />

in this case, they are also clients) like the voluntary sector organisations, private organisations etc. The<br />

government provides the resources in terms of equipment, funds, facilities and personnel needed to<br />

deliver the services to the patients. The government in the process of resourcing the NHS sets targets<br />

through the Department of Health for health and social service providers. But for the health care system<br />

to be effective in meeting the local needs and expectations, it has to include the inputs of the clients and<br />

patients. According to some clients (patients) interviewed in Lincolnshire in England, the system is not<br />

performing because some of their colleagues had to go back to the hospital after their medical treatment<br />

because the ‘system’ has no adequate provision for social integration and social needs of the patients. It<br />

is argued that, where such provisions exist, the government decide what is needed and the criteria for<br />

measuring and managing the performance without considering the local needs of the citizens which may<br />

be different from one locality to another. The clients (patients) claimed that where the system seems to<br />

give attention to their needs, the government thinks the system is not performing because it is not<br />

meeting the targets set and therefore closes down such hospitals or departments in the hospitals even<br />

when the citizens protest.<br />

It is being argued that mainstream mental health care researches tend to be clinically dominated,<br />

reflecting the linear and hierarchical approaches of medical professionals who based their analysis on<br />

quantitative and experimental testing with less attention to qualitative research. One of the focuses of the<br />

paper therefore is to give attention to qualitative research in mental health services by analysing how to<br />

give audience to and manage service users, carers, the family and the community’s inputs through the<br />

actors-dialogism-system.<br />

1


2. Mental health care services in UK<br />

Kwesi Korsa Aggrey<br />

Mental health problems account for significant proportions of diseases in UK. However, access to<br />

treatment though available is not being fully utilised because of a number of issues in which<br />

reorganisation of the service is urgently needed. The issues of mental health care have become<br />

important because of its significant burden on the government and the community. It is argued that, it<br />

accounts for nearly 12% of the global burden of disease and will hit the target of 15% of disability<br />

adjusted life years lost to illness by 2020 (WHO, 2003).<br />

The economic and social costs to the state and the community have increased with the development of<br />

science and medicine. Before the 17 th century, issues of mental disorders were attributed to spiritual<br />

explanation and therefore its treatment. However, with the development in science and medicine, secular<br />

explanations were given to the mental disorder as a physical state of the mind leading to the confinement<br />

of such people in secured places.<br />

During the first part of the 18 th century however, the dominant view of mental disorders being incurable<br />

and the justification of being sub-humans and therefore their status, living conditions and physical<br />

restraints to places of confinement changed.<br />

The pressure from society and humanitarian groups about the human right of people with mental<br />

disorders led to a call for reforms in the treatment of mental patients and therefore the introduction of<br />

moral treatment programmes. This led to deinstitutionalisation or de-hospitalisation of the service delivery<br />

aimed at creating the opportunity to implement some of the ideas developed on social network and<br />

community support system as alternatives outside psychiatric institutions.<br />

However, the way mental health care service delivery is organised determines the effectiveness of the<br />

interventions and the ultimate fulfilment of the objective of the service. The quality of the service<br />

determines the likelihood of one achieving the desired outcome of becoming employable, independent,<br />

self organised and part of the community. The quality of the service delivery is important because it<br />

ensures that service users receive the care and support they require; it also ensures that there is<br />

improvement in their daily life activities and living; provide acceptable and relevant clinical and non<br />

clinical care to reduce the impact of mental health problems; and to make efficient and effective use of<br />

resources available to them (Murray and Lopez, 1996).<br />

In UK, the government’s policy on community care services was strongly developed with the third sector<br />

in mind. This is evidenced in the Community Care Act (Department of Health, 1990) which embraces the<br />

full involvement of other actors in mental health care services in particularly, the Third Sector. Despite<br />

this policy however, the collaboration between the statutory and the actors particularly the voluntary<br />

sector organisations to improve the quality of the service is weak (Simpson, 1996) and problematic,<br />

(Adams, 1990). The weakness of this collaboration has been attributed to lack of understanding of each<br />

other’s objectives, missions or vision and organisational practices, methods and significantly, differences<br />

in respect of their structures, resources (Wilson, 1994) and their value systems (Mosher and Burti, 1994).<br />

There is also an issue of confidence and trust by the professionals of the formal sector about the<br />

capabilities of semi skilled personnel of the Third Sector and the private sector. Gratham (1995) surveyed<br />

55 general practitioners in his area and found that only 18 per cent had made some referral to the<br />

Alzheimer’s disease Society. Some professional even see any sharing of skills, knowledge and<br />

experiences with the non-professional as a threat to their profession (Gussow and Tracy, 1976). The<br />

paper aims at proposing a method of resolving the problem of collaboration among the actors through<br />

participatory communication (dialogism). We will now look at the methodology used in this research to<br />

identify the problem of limited communication among actors of the system.<br />

3. Methodology and design<br />

The study is intended to expand on our understanding of how to improve performance or improve the<br />

quality service through increase of interaction among the stakeholders in health service delivery,<br />

particularly in mental health care delivery system. A phenomenological method was used in this research<br />

because it is one of the methods which help to identify phenomena through the perspectives of the actors<br />

in the situation. In human interactive situations like the mental health care service delivery, it enables the<br />

researcher to gather in depth information and perceptions through qualitative methods of interviewing and<br />

focal group discussions (Gerber, 2001).<br />

2


Kwesi Korsa Aggrey<br />

Phenomenology could be looked at as a philosophy, but also as a research method for gathering and<br />

analysing data (Goulding, 1999). As a method, phenomenology allows for the discovering of people’s<br />

‘world views’ thus capturing their subjective experiences. In all, eighteen (18) managers from the<br />

Lincolnshire Partnership Foundation Trust (LPFT), Lincoln County Council (LCC) Lincoln County Hospital<br />

(LCH) and some voluntary sector organisations in Lincolnshire were interviewed. In addition two focal<br />

groups’ discussions were also conducted.<br />

Using the phenomenological methods of interviews and focal groups discussions, the experiences of the<br />

participants were organised by means of storytelling and interwove with living story method. The living<br />

story theory, which is the bedrock of living story method is defined as the emergence, trajectory, and<br />

morphing of living story from ante-narrative-conception to the end of decomposition of a story (Boje,<br />

2007b; Boje, 2008). A living story is a story which lies in-between dead and alive narrative and between<br />

forgotten and revitalised story. It is therefore a critical ante-narrative, where a living story traces and predeconstructs<br />

an ongoing interweaving living story narrative and ante-narrative to become selfdeconstruct.<br />

In the process of deconstructing stories, some stories die and others rejuvenate or emerge<br />

and better still some revolve. The living story method is therefore interplay of dead (ended) narratives<br />

with living emergent stories. It tells the emergent stories in the context of fragmented dead narratives.<br />

The data collected were transcribed and analysed by reading and analysing the data in depth to search<br />

for differences and patterns in relation to how to improve service delivery. Each interview was used to<br />

probe into issues from the previous interviews which were not clear, in order to create a ‘rich picture’ of<br />

the situation (lack of adequate communication among the actors). The data were interpreted,<br />

continuously revised and the context broadened using the notion of ‘Ideal interpreter’.<br />

Interpretivism is assumed to be mental content judgement-dependent; that is, the facts about<br />

propositional attitude of people is exactly what is being captured by the judgement of the Ideal Interpreter<br />

(Johnston, 1993a). However, for Dennett and Davidson (sited in Wright, 1989), the Ideal Interpreter is a<br />

third-person who interpreters someone else. The interpreter can also be a first- personal account where<br />

the interpreter is taken to be the subject of interpretation. Dennett’s and Davidson’s version of interpreter<br />

is however, formulated on the thesis of biconditional: that X belief Q, therefore if there is an informed<br />

Ideal Interpreter, the Ideal Interpreter would be disposed to attribute to X the belief that is Q. There is<br />

however, a problem with this line of argument of interpretivism. There is no a priori guarantee that the<br />

Ideal Interpreter will find answers to all the meaning of the subjects’ beliefs particularly those that are<br />

linguistically expressed. It is therefore necessary that some sufficient conditions are held as a constraint<br />

to strengthen this identified weakness of interpretivism, that is, there is an appropriate informed Ideal<br />

Interpreter who would be disposed to attribute X the belief that Q. Interpretivism presupposes therefore<br />

that facts or believes given by the Ideal Interpreter will be the facts participants believe in an ideal<br />

situation, that is the assumption.<br />

For the purposes of this research, the Ideal Interpreter is assumed to have a sufficient database of nonintentional<br />

and intentional sources of information, knowledge and experience to interpret the subjects’<br />

meanings, believes, experiences and desires. The Ideal Interpreter uses daily life application of intentions<br />

in addition to scientific minded observations and experiences of non-intentional in order to arrive at the<br />

best possible meanings of the participants’ experiences and believes in order to detach the meanings of<br />

the participants’ experiences at the individual level. The next section looks at how interaction among the<br />

actors through participatory communication will help to achieve the ultimate level of collaboration that will<br />

encourage improvement in the quality of the service delivery system.<br />

4. Actors’ collaboration: Way forward to improve the quality of service<br />

It is proposed that to improve the quality of service, collaboration among all interested parties of the<br />

mental health care delivery system is necessary in order to meet the needs and expectations of clients.<br />

One can say therefore that, the collaboration among the stakeholders is not happening because of<br />

communication problems among the actors of the service delivery system. This was identified and<br />

confirmed from the data collected and organised as per the emerged sub themes (namely: state control<br />

of the mental healthcare system; organisational culture; service delivery as continuous process; service<br />

users needs and expectations; and participatory communication) from interviews conducted among<br />

managers and directors from the Lincolnshire Partnership Foundation Trust (LPFT), Lincoln County<br />

Hospital (LCH), Lincoln County Council (LCC),some voluntary sector organisations and some service<br />

users from the county, (these are some of the actors of the mental health care system).<br />

There are a number of possibilities to resolve the problem identified:<br />

3


Kwesi Korsa Aggrey<br />

One possibility is to focus on information overload in the organisation of the NHS. It is proposed, by<br />

increasing the level of interaction among all the interested parties via the use of participatory<br />

communication after a suitable destabilisation of the existing communication system, the quality of<br />

service will be improved.<br />

Another possibility is the proposal to use the method of ‘information overload’, i.e. an intentional increase<br />

in information through the communication channels of the NHS, as a way to support a suitable<br />

destabilisation of the interaction among all the interested parties to stimulate self-organised improvement.<br />

The third possibility, the one that is the paper’s focus is, the possible improvement in the organisation of<br />

the mental health care system within the NHS. It is proposed therefore that the use of ‘information<br />

overload’, a process to achieve participatory communication is necessary in order to increase the level of<br />

interaction among all the interested parties and also to stimulate self-organisation of the system using the<br />

dialogism approach. It is when the level of interaction in the decision making process is increased and<br />

more attention given to the beneficiaries of the system through the supporting groups’ representatives<br />

that the system may attain optimum performance and meet clients’ needs, expectations and in fact satisfy<br />

the needs of all stakeholders, that is achieve ‘complete collective satisfaction’.<br />

One theme that emerged from the interviews conducted was the organisational culture of the NHS, which<br />

mainly is about professionalism of work, staff and the way task is undertaken or service is rendered. The<br />

NHS: PCT and LPFT staffs have the idea that work should be based on specialisation of tasks leading to<br />

the departmentalisation and training of staff accordingly.<br />

To deal with the culture of professionalism and also to improve the services delivery, there is the need for<br />

collaboration, because after medication and discharge from hospital, service users may need the social<br />

support mainly provided by other actors including the Third Sector. It is a fact that the majority of mental<br />

health patients preferred to work with and receive help from the non-professionals than from the mental<br />

health professionals (Barker et al, 1990). Research has proven that, for mental health problems in<br />

particular, the benefit of social support cannot be swept under the carpet especially at the time of<br />

discharge from hospitals and that social networks has been very influential in the improvement of quality<br />

of life of service users. (Green et al, 2002).<br />

It is therefore important that collaboration is encouraged among all stakeholders to achieve complete<br />

collective satisfaction. This is necessary because no single organisation can meet the needs and<br />

expectations of mental health service users on its own, therefore a multi-agency model (actors-dialogismsystem)<br />

should be explored via collaboration; training of self help groups, sharing information, collective<br />

service evaluation and feedback to improve services. It is also important to incorporate other methods<br />

used by the Third Sector staff to strengthen the collaboration between the third sector and the formal<br />

sector (Simpson, 1996).<br />

It is important to note that in order to evaluate or manage the performance of the service delivery in<br />

mental health care and to improve the quality, one needs to go beyond the medical model, diagnosis of<br />

symptoms, treatment and side effects of disease. This may involve input/output system performance<br />

evaluation. This system of input/output evaluation is what the researcher calls the First Cybernetics (Boje,<br />

2008). However, for total treatment of mental problems, the inclusion of social well being of patients<br />

(service users) is necessary. This in a collective term is referred to as Health-Related Quality Of Life<br />

(HRQOL). The HRQOL generally includes the domains of physical functioning, psychological, well being<br />

(e.g. level of anxiety, depression, fear of recurrence, etc); and social functioning of the patients. The<br />

process of evaluating the performance of the mental service delivery system goes beyond the medical<br />

model, input/output approach, to include the system’s environment (open system approach).<br />

Cybernetics at the level of input/output approach is about negative feedback within self-stabilising loops<br />

that lead to a regulation of a system, in order to generate a state of equilibrium (Ashby 1957). When<br />

Cybernetics is related to external and centralised control of black boxes and goal-oriented behaviours, in<br />

order to sustain complex interactions with the environment over time (Beer, 1966 p.257), then, the<br />

system turns to become ‘open system’ a Second Cybernetics system or approach (Boje, 2008).<br />

Cybernetics might also manifest itself in connected networks of components in interaction. As a result of<br />

such interconnectedness, multiple feedback loops may be formed within the networks, which allow for<br />

self-regulation and the production of ‘global’ order. This implies that a system can also have distributed<br />

4


Kwesi Korsa Aggrey<br />

control, which does not depend on an external centralised control unit (Ashby, 1957), this is termed the<br />

First Cybernetics.<br />

The concept of negative feedback is crucial for achieving stability band adaptation. Negative feedback is<br />

a necessary corrective action, which when implemented produces a specific outcome by modifying a<br />

course of action. The consequence of this arrangement is that, the input of the loop is affected by its own<br />

output. Hence, a satisfactory outcome can be achieved following a negative feedback process, indicating<br />

whether the system has achieved its goal or fallen short. It does this by providing backward information<br />

on the resulting outputs, in order to be used for the manipulation of the inputs of the system (Ashby 1957<br />

p.53-54). This is what is expected from a participatory communication process via dialogism of actors’<br />

interaction if fully implemented.<br />

The First Cybernetic approach as described above is not very suitable for a complex system like the<br />

mental health care system. The reason being that as adaptation and stability is achieved; the system is<br />

bound to dissipate (dysfunction) with time because of the continuous changes and constant increase of<br />

the variety in its environment, which may exceed the capacity of the input/output cybernetics system. This<br />

is in line with the reverse interpretation of the Ashby’s Law of requisite variety. Therefore one needed to<br />

go beyond this level of abstraction to the Second Cybernetics system which is the open system thinking.<br />

It is an open system because it receives its feedback not only from within the system but also from the<br />

system’s environment. Again being an open system, it means the system does not only exchange energy<br />

with its environment but also matter from the environment. However, in both the First and Second<br />

Cybernetics approaches of improving performance or quality of service, there are elements of linearity<br />

and hierarchical, which in themselves create problems<br />

The mental health care system, as a human institution with different people with different needs and<br />

preferences, it is expected that the capacity of its meta-system transition (co-ordinated process) will not<br />

be able to cope with the variety of needs and preferences from both inside and outside the system and<br />

therefore dissipation is bound still to happen. This is what has resulted in various forms of reforms all<br />

aimed at resolving the problems associated with increase variety of needs and preferences by the<br />

various actors. One cannot avoid the dissipation because all the actors have their own different goaloriented<br />

values which are not static but grow together with a diversifying environment, which may be<br />

different from government’s target. The only way to resolve the problem in order to achieve the desired<br />

high quality of service delivery (complete collective satisfaction) is to move to a higher level of cooperation<br />

among the actors where the relationship between them or the functional imperative is neither a<br />

linear nor hierarchical but spiral relationship of polypi dialogism. This is an inter-co-ordinated system<br />

function of various dialogisms, namely the polyphonic, stylistic, chronotonic and architectonic.<br />

5. Dialogism<br />

A polyphonic is a written, visualised or orally told stories by all the stakeholders of a system as opposed<br />

to mono vocal narrative or written strategy by an expert or a dominant actor (Boje, 2008). It is a<br />

construction by many embodied voices, logics and perceptions. It could be visual arts, photos, décors,<br />

drama or oral storytelling that communicates through interaction. It involves getting stakeholders to<br />

engage in storytelling through the sharing of ideas, experiences, knowledge, skills such that each and<br />

every one’s voice and logic get enunciated and subjected to questioning until a form of understanding is<br />

reached on the particular issue and how to deal with it.<br />

However, the interaction or communication in the process of the sharing is not necessarily a function of<br />

producing a consensus but usually results in open situation where the issue is being accepted or<br />

rejected. The actors in the process therefore reach a point, which bifurcates into further alternatives. The<br />

process of bifurcation itself is a process of reducing complexity of the issues that come to the ‘floor’ of the<br />

dialogue into simple terms and meanings.<br />

The Stylistic is a dialogue; it is an orchestration of image or dialogism among various means of dialogue<br />

or communication namely oral telling of stories, print and video media, internet, gesture-theatrics, décor<br />

and other architecture modes of image expressions. It is a juxtaposition of varied styles for interaction. It<br />

can also be described as the interactivity of various modes of expressions of organisational image.<br />

For Weick (1995), sense making could be a way of looking at stylistic on the basis of public sense making<br />

control. Sense making is about the act of or the process of placement of items into frameworks,<br />

comprehending, redressing and constructing meanings from the frame such that interacting with it<br />

5


Kwesi Korsa Aggrey<br />

generate mutual understanding and patterns. This form of framing and reframing (producing and<br />

reproducing) is a stylistic feature. It also determines the style of delivery of the services and where<br />

necessary the revision of the style to meet certain moral obligations. The power of stylistics is in the way<br />

that stylistics is able to connect or depend on the working interplay of the roles of all stakeholders in the<br />

service delivery to achieve a common purpose.<br />

Chronotonic is a relativity of time and space, in terms of functions; it is like Einstein’s theory of relativity,<br />

with time being the fourth dimension of space. Chronotopic dialogism can also be described as a<br />

holographic relation of centring (centripetal of chronotopes) and amplifying (centrifugal of chronotopes).<br />

Holographic is storytelling that runs from one dimension (monogon) to multiple complexities. Storytelling<br />

is holographic in the sense that it can interrelate to more than one complexity. Storytelling is infectious<br />

and that can be tracked across space-time.<br />

Architectonic is the orchestration of ethics in relation to aesthetic and cognitive storytelling. Architectonic<br />

dialogism is mainly focus on the interaction of several societal discourses that affects organisation’s<br />

performance. The three basic discourses are ethics, aesthetic and cognitive which are answerable to one<br />

another. Kant (1993) and Bakhtin (1981) described architectonic as societal discourses. However, each<br />

author has his own divergent views on this notion. Kant invented the ‘cognitive architectonic’. For him<br />

architectonic is the art of constructing a system. He argues that reason cannot permit our knowledge to<br />

remain in an unconnected and rhapsodistic state, but requires that the sum of one’s cognitions constitute<br />

a system. Kant sees architectonic as cognitive notion deeply implicated in the construction of systemicity.<br />

Bakhtin preferred the term ‘consummation’ to construction and was careful not to assume a monophonic<br />

or mono logic or mono language system. He looked at a system as ‘systemicity’, that is unmergedness<br />

and unfinalisability of a system. This is the way forward for the proposed mental health care system,<br />

where there is no room for homeostasis state or dissipation because the process of ‘polypi’ is<br />

‘unmergedness’ ‘and unfinalisation’. Bakhtin went further from Kant’s cognitive architectonic discourse to<br />

add ethical and aesthetic discourses. He defined Ethics discourse as the notion of ‘answerability’ which is<br />

a description of how one domain of discourse is answerable to another. The Aesthetic is about how and<br />

for whom a given systemicity is constructed or designed.<br />

Polypi is the inter-dialogism of polyphonic, stylistic, chronotopic and architectonic. It is a multi-dialogised<br />

complexity where the four (4) notions collide. The polypi is therefore a coordination system of the<br />

dialogism<br />

6. Conclusion<br />

The complexity of dialogism, the polypi, which explores multiple of dialogisms, is in line with Letiches’s<br />

(2000) phenomenal complexity and also Stacey’s (2006) emergent complexity. The properties of the<br />

complexity are cumulative rather than successive. In cumulative, all the lower orders of phenomenal<br />

complexity (polyphonoic dialogism) intermingle with higher orders of dialogism (of fragmented antenarrative<br />

and petrified narratives that follow Aristotelian structures of wholeness and coherence of<br />

beginning middle and end in storytelling.<br />

Weick (1995) looks at the complexity in retrospective sense making of experiences and narrative of<br />

coherence and control by people whose current experiences fit into past meaning structures. This I think<br />

offers a revolutionary breakthrough in complexity thinking or approach .This approach is not only<br />

compatible with phenomenal complexity theory, but also provides a way to overcome a major crack or<br />

gap in knowledge creation. The same argument applies to synchronous and diachronous approaches to<br />

storytelling organisation. The synchronous approach (being the current experienced story) looks at<br />

storytelling organisation at a particular point in time, rather than over time while the diachronous<br />

approach (being narrative) looks at storytelling organisations in historical development. Therefore in order<br />

to achieve the wholeness, one has to bring both approaches together through the phenomena, the<br />

‘polypi’ which is the dialogism of history or background of current issue(s) plus current experiences of<br />

users of the mental health care delivery system.<br />

A typical and recent familiar example of narrative and emergent stories of dialogism is the death of<br />

Michael Jackson, the pop star. A network of organisations constructed narratives along side stories of<br />

Michael Jackson’s death; even now stories about his death continue to emerge. In the narrative and<br />

emergent stories of his death, the polypi of dialogism is contentious: polyphonic logics struggle to<br />

converge or agree; the multiple stylistics of verbal, written and posters or pictures contrast; the multiple<br />

6


Kwesi Korsa Aggrey<br />

chronotopes of varying temporalities and partialities diverge, and the ethical discourse of architectonic<br />

questions reverberates into many other discourses. It indicates therefore that the complexity of the<br />

narratives and emergent stories of Michael Jackson’s death is unmerged and ‘unfinalised’ if one want to<br />

get to the ‘bottom’of the story.<br />

Dialogism is a complex notion, it is a weave of many storytellers and listeners who together co-construct<br />

the meanings of a dynamic that reduce living story into complete collective meaning as a necessary<br />

outcome as opposed to ante-narrative amplification (Boje, 2008).<br />

The contribution of this paper to knowledge is the proposal that although linearity or hierarchical<br />

approaches to knowledge creation is making impact in society, the impact will be great and more<br />

beneficially however, if the interaction among the actors is on a spiral relationship, that is on a collective<br />

level by way of collaboration to resolve the problems identified collectively in order to achieve ‘complete<br />

collective satisfaction’ particularly in the service delivery system.<br />

In the case of mental health care delivery system ‘Complete collective satisfaction’ is achievable only if<br />

the mental health care system moves from the level of abstract categorisation such as planning or<br />

reforms to the level of collaboration. This means that the way forward is therefore, the re-organisation of<br />

the system which is exclusive to the NHS or government (target setting) to a level of multifunctional and<br />

interdependent interaction where consideration is given the environment, other service providers, clients,<br />

financiers, the public and in fact all stakeholders.<br />

References<br />

Adams, R. (1990), Self-help, Social work and Employment. Macmillan, London.<br />

Ashby, W. R. (1957) An Introduction to Cybernetics, Chapman & Hall, London<br />

Ashby, W. R. (1958) “Requisite variety and its implications for the control of complex systems” Cybernetics, Vol. 1<br />

No.2 pp83-99<br />

Ashby, W.R (1980) An Introduction to Cybernetics. London: Chapman and Hall L<br />

Bakhtin, M. M. (1981) Art and Answerability (Michael Holquist and Vadim Liapunor, eds; translation and notes by<br />

Vadim Liapimov; supplement translated by Kenneth Brostrom) Austin TX: University of Texas Press<br />

Barker, et al (1990) “Coping and help seeking in the UK adult population” British Journal of Clinical Psychology, Vol.<br />

29, pp.271-85<br />

Beer, S. (1966) Decision and Control: The Meanings of Operational Research and Management Cybernetics,<br />

London: Wiley.<br />

Beer, S. (1979). The Heart of Enterprise Chichester: Wiley<br />

Beer, S, (1985) Diagnosing the System for Organisations. John Wiley, Chichester<br />

Boje, D. M. (2007b) From Wilda to Disney: Living Stories in Family and Organisation Research. In Jean Clanddinin<br />

(ed.) Handbook of Narrative Inquiry, London: Sage<br />

Boje, M. D. (2008) Storytelling Organisations Sage Publications Limited London<br />

Department of Health (1990), National Health Service and Community Care Act, HMSO, London<br />

Donaldson, M. S. (2004) “Taking Stock of Health-Related Quality-of- Life Measurement in Oncology Practice in the<br />

United States” JNCI Monographs<br />

Vol. 2004 No. 33 pp. 155-167 Oxford University Press London<br />

Foucault, M. (1977b) Language, Counter-Memory, Practice: Selected Essays and Interviews by Michel Foucault, D.<br />

F. Bouchard (ed.). Trans. and intro Ithaca, NY: Cornell University Press.<br />

Gerber, R. (2001), “The concept of commonsense in workplace learning and experience”<br />

Goulding, C. (1999) “Consumer research, interpretive paradigms and methodological ambiguities”, <strong>European</strong> Journal<br />

of Marketing, Vol. 33 No. 9/10, pp. 859-73<br />

Graham, n. (1995) “GPs and voluntary organisations” British Journal of General Practice, Vol. 45, p. 273<br />

Green, G. et al (2002) “The role and impact of social relationships upon well being reported by mental health serviceusers:<br />

a qualitative study”, Journal of Mental Health, Vol. 11, pp.565-79.<br />

Gussow, Z.and Tracy, G.S. (1976), “The role of self-help clubs in adaptation to chronic illness and disability”, Social<br />

Science and Medicine, Vol. 10, pp. 407-14<br />

Johnston, M. (1993a) “Objectivity Refigure: Pragmatism without Verification” in J. Haldone and C. Wright, eds.<br />

Reality, Representation and Projection Oxford University Press.<br />

Kendall, J. and Knapp, M. (2000)” Measuring the Performance of Voluntary Organisations” Public Management 2(1):<br />

105-132<br />

Letiche, H. (2000) Phenomenal complexity theory as informed by Bergson. Journal of Organisational Changes<br />

Management, 13 (6): 545-58.<br />

Mosher, L.R. and Burti, L. (1994) Community Mental Health: Principles and Practice, New York, NY<br />

Murray, A. and Shepherd, G. (1996) “Suspicious minds”, Health Services Journal, Vol. 106, p. 27.<br />

Simpson, R. G. (1996) “Relationship between self-help organisations and professional health-care providers”, Health<br />

and Social Care in the Community, vol. 4, pp. 359-70<br />

7


Kwesi Korsa Aggrey<br />

Stacey, R. (2006) Complex responsive processes as a theory of organisational improvisation, pp128-41 London:<br />

Routledge<br />

Weick, K. E. (1995), Sense making in Organisations, Sage, Thousands Oaks, CA<br />

Wilson, J. (1994) “Aspirations and realities: community care at the crossroads”, Health and Social Care in the<br />

Community Vol.3, PP. 227-40<br />

World Health Organisation (WHO), Mental Health in the WHO <strong>European</strong> Region, Fact sheet EURO/03/03,<br />

Copenhagen, Vienna, 2003.<br />

Wright, C. (1989) “Wittgestein’s RuleFollowing Considerations and the Central Project of Theoretical Linguistics” in A.<br />

George, ed. Reflections on Chomsky, Basil Blackwell<br />

8


The Acceptance and use of Educational Technology in<br />

Kingdom of Bahrain<br />

Jaflah Al-Ammary<br />

University of Bahrain, Bahrain<br />

Jaflah@itc.uob.bh<br />

Abstract: In 2004, King Hamad Schools of Future Project (KHSFP) has been launched by the ministry of education<br />

at Kingdom of Bahrain with a goal of establishing a fundamental change in the way teaching and learning take place.<br />

The main objective of the project is to change the way in which students and teachers communicate, work and learn<br />

by transfer the traditional classroom into an open, interactive environment based on adopting new and more advance<br />

technologies. In fact at Kingdom of Bahrain, adopting technologies such as computers, smart board, or interactive<br />

board will increase the pressure on the schools to accommodate to these new conditions. This is because there are<br />

some schools in particular the elementary schools were only getting access to the technology and many teachers<br />

lack the required IT skills and capabilities. As a result, many efforts have been initiated to enhance the necessary<br />

skills and facilitate the functional development of the teachers. However, there is still a need to understand the<br />

teachers and student’s acceptance of the new technology to cope with KHSFP. Thus, to achieve a technological<br />

enhancement of the education system, students and teachers should accept the IT as an effective tool. The current<br />

research aims at investigating the acceptance and use of the new educational technology such as computer,<br />

interactive board and smart board, by teachers and students in the schools of future at Kingdom of Bahrain. Through<br />

an extension of the Technology Acceptance Model (TAM), five factors that influence the behavioral intention to<br />

accept and use education technology were examined. These factors include: subjective norms, motivation, computer<br />

self efficacy, previous experiences, and computer anxiety. Moreover, previous experiences was investigated to has<br />

an indirect effect on perceive ease of use through the computer anxiety. The results show that the computer self<br />

efficacy, previous experience, and motivation are the most critical factor for accepting and use educational<br />

technology at Kingdom of Bahrain. By considering these factors and investigating other factors, Ministry of Education<br />

at Kingdom of Bahrain can enhance the acceptance and use of the new educational technology by both teachers and<br />

students among the Kingdom schools.<br />

Keywords: acceptance technology model, computer anxiety, educational technology and previous experience<br />

1. Introduction<br />

In order to get benefit from the opportunities offered by the digital revolution and the big technological<br />

innovations in the field of education; the ministry of education at Kingdom of Bahrain decided to draw a<br />

strategy to employ ICT in education (Shaker, 2008). This strategy aims at developing a new environment<br />

for e-learning, not just providing computer laboratories with equipments and teaching computer as<br />

subject matter (ICE47, 2001). Having such action, future generations will be prepared to establish<br />

information society and be empowered with the basic skills necessary to transfer the kingdom into a<br />

knowledge based economy (Shaker, 2008, ICE47, 2001). The ministry of education is making its efforts<br />

to achieve its objective through the implementation of big projects that deal mainly with enhancing and<br />

improving the quality of education at Kingdom of Bahrain (ICE47, 2001). One of these projects is KHSFP.<br />

In 2004, King Hamad Schools of Future Project (KHSFP) has been launched by the ministry of education<br />

at Kingdom of Bahrain with a goal of establishing a fundamental change in the way teaching and learning<br />

take place. The main objective of the project is to change the way in which students and teachers<br />

communicate, work and learn by transfer the traditional classroom into an open, interactive environment<br />

based on adopting new and more advance technologies. (Al-Ammary et al. 2010a,b). Moreover, the<br />

project will reform the education system in Kingdom of Bahrain to prepare the citizen to invest in the ICT’s<br />

potentials to achieve the education quality and attain the curricula capabilities at all educational levels<br />

(ICE47, 2001). KHSFP includes an integrated systems comprise of an educational portal to achieve the<br />

qualitative turning point in the educational performance under a contemporary educational environment,<br />

supported by a modern educational and information technology (ICE47, 2001). This project has been<br />

implemented by first, establishing a website for the project, second, disseminating information about the<br />

project in the target schools and third, establishing a scientific documentary library for the project<br />

(ESCWA, 2005). The project will be executed by connecting the targeted schools (primary, elementary,<br />

and secondary level), with a speedy communication network via the central educational portal (ESCWA,<br />

2005).<br />

In fact at Kingdom of Bahrain, adopting technologies such as computers, smart board, or interactive<br />

board will increase the pressure on the schools to accommodate to these new conditions. This is<br />

because there are some schools in particular the elementary schools were only getting access to the<br />

9


Jaflah Al-Ammary<br />

technology and many teachers lack the required IT skills and capabilities. As a result, many efforts have<br />

been initiated to enhance the necessary skills and facilitate the functional development of the teachers.<br />

However, there is still a need to understand the teachers and student’s acceptance of the new technology<br />

to cope with KHSFP. Thus, to achieve a technological enhancement of the education system, students<br />

and teachers should accept the IT as an effective tool. Therefore, the current research has conducted to<br />

investigate the acceptance and use of the new educational technology such as computer, interactive<br />

board and smart board, by teachers and students in the schools of future at Kingdom of Bahrain. Through<br />

an extension of the Technology Acceptance Model (TAM), five factors which influence the behavioral<br />

intention to accept and use education technology were examined. These factors include: subjective<br />

norms, computer self efficacy, previous experiences, motivation and computer anxiety. Moreover,<br />

previous experiences was investigated to has an indirect effect on perceive ease of use through the<br />

computer anxiety.<br />

This paper consists of nine sections; each section will touch a significant component of this research. The<br />

following section spots the light on the literature review on the Technology Acceptance Model (TAM).<br />

Section three explains the research model and hypotheses. The survey instruments of this study are<br />

discussed in section four. Section five illustrates the research methodology for this study. Description of<br />

data analysis is presented in section six. In section seven, profile of the study participants are discussed.<br />

Finally, in the last two sections of this paper; 8 and 9, the findings of this study are discussed and a<br />

conclusion is presented and made ready for the educational institutes that intend to adopt IT.<br />

2. Research background<br />

2.1 Technology Acceptance Model<br />

There are many models have been developed to investigate and understand the factors affecting the<br />

acceptance of computer technology in the organization. Among those are the Theory of Reasoned Action<br />

(TRA) (Ajzen et al. 1980), Theory of Planned Behavior (TPB) (Ajzen, 1991), Technology Acceptance<br />

Model (TAM) (Davis et al., 1989), and Unified Theory of Acceptance and Use of Technology (UTAUT)<br />

(Venkatesh et al., 2003). However, it have been noticed that the most used model by Information System<br />

academics and practitioners is the TAM designed by (Davis, 1998). TAM is an adaptation of the Theory<br />

of Reasonal Action (TRA) to the field of IS. The objective of TAM is to provide an explanation of the<br />

determinants of the adoption and use of information technology (Davis et al., 1989). Perceived<br />

usefulness is the degree to which a person believes that using a particular system enhances his or her<br />

job performance. Perceived ease of use is the degree to which a person believes that using a particular<br />

system will be free of effort (Davis et al., 1989). Perceived ease of use and perceived usefulness are<br />

common in technology-usage setting (Tylor et al., 1995 a, b) and have been applied widely to solve the<br />

acceptance problem using different acceptance models (Wu et al., 2005).<br />

TAM originally suggested that perceived usefulness and perceived ease of use are involved in explaining<br />

the variance in user’s intention or behavior of intention (Davis et al., 1989). Perceived usefulness is also<br />

seen as being directly impacted by perceived ease of use. However, behavior intention to use functioning<br />

has hypothesized by TAM as a mediator of actual technology in use (Davis, 1998).<br />

3. Research model and hypotheses<br />

This research study has a main objective of investigating factors that influence the acceptance of<br />

educational technology in the schools at Kingdom of Bahrain. The literature revealed for many factors<br />

that may affect the acceptance and use of educational technology in any school. For the purpose of the<br />

current research, five factors have been selected which include: subjective norms, previous experience,<br />

computer anxiety, computer self efficacy, and motivation. The research model is developed as depicted<br />

in Figure (1).<br />

3.1 Theory factors: Perceived usefulness and perceive ease of use<br />

Information system researchers have investigated TAM and agreed on its validity in predicting the<br />

individual's acceptance of various systems (Chin and Todd, 1995; Doll et al, 1998; Segars and Grover,<br />

1993; Venkatesh and Davis, 2000). Most of them revealed that perceived usefulness and perceived ease<br />

of use have a positive effect on the behavioral intention to use system (Chin and Todd, 1995; Doll et al,<br />

1998). Moreover, perceived ease of use was found to have a positive effect on the perceived usefulness<br />

(Lee, 2006). The hypotheses relating to perceived ease of use and perceived usefulness have been<br />

studied by a number of TAM researchers (Srite, 2006). Although some researchers reveled that attitude<br />

10


Jaflah Al-Ammary<br />

toward systems use is postulated to mediate the effect of perceived ease of use and perceived<br />

usefulness on behavioral intention (Kim, el al, 2009), Davis, et al, (1989) suggest that perceived<br />

usefulness may impact on behavioral intention to use technology irrespective of attitude toward this<br />

system.<br />

Computer<br />

Self efficacy<br />

Motivation<br />

Previous<br />

experiences<br />

H12<br />

Computer<br />

Anxiety<br />

H<br />

H7<br />

H9<br />

H8<br />

H6<br />

H10<br />

H1<br />

H13<br />

H14<br />

H5<br />

Subjective<br />

Norms<br />

Perceived<br />

usefulness<br />

experiences<br />

H<br />

Perceived<br />

ease of use<br />

H<br />

H<br />

H<br />

Behavior<br />

Intention<br />

Figure 1: Research model<br />

Hypothesis 1: Perceived usefulness has a positive effect on students and teacher’s behavioral intention<br />

to use educational technology in school at Kingdom of Bahrain.<br />

Hypothesis 2: Perceived ease of use has a positive effect on students and teacher’s behavioral intention<br />

to use educational technology in school at Kingdom of Bahrain.<br />

Hypothesis 3: Perceived ease of use has a positive effect on students and teacher’s perceived<br />

usefulness of educational technology in school at Kingdom of Bahrain.<br />

3.2 Hypotheses on factors affecting the perceived ease of use and perceived usefulness<br />

3.3 Subjective norms<br />

According to Fishbein and Ajzen (1975), subjective norms refer to “the person’s perception that the most<br />

people who are important to him think he should or should not perform the behavior in question”. The<br />

literature reported on much theoretical and empirical evidence regarding the importance of the role of<br />

subjective norms on technology use directly, or indirectly through perceived usefulness in the workplace<br />

(Lee, 2006; Tylor et al., 1995a). Subjective norms have been found to be more important prior to, or in<br />

the early stages of innovation implementation when users have limited direct experience from which to<br />

develop attitude (Havelka, 2003). Venkatesh and Davis (2000) found in their study that there is a direct<br />

effect of subjective norms on the behavioral intention to use information systems. Lee [24] reported that<br />

the effect of subjective norms has significantly influenced perceived usefulness. Accordingly, the<br />

following hypothesis was developed accordingly:<br />

Hypothesis 4: Subjective norms have a positive effect on students and teacher’s behavioral intention to<br />

use educational technology in school at Kingdom of Bahrain.<br />

Hypothesis 5: Subjective norms have a positive effect on students and teacher’s perceived usefulness of<br />

educational technology in school at Kingdom of Bahrain.<br />

11


3.3.1 Computer self efficacy<br />

Jaflah Al-Ammary<br />

Computer self efficacy is defined as individual’s self confidence in his or her ability to perform behavior<br />

and use the computer in the context of information technology usage (Compeau et al., 1995; Compeau et<br />

al., 1999). According to Hayashe, et al. (2003), the computer self-efficacy is not concerned with what one<br />

has done in the past, but rather with judgments of what could be done in the future. The effect of<br />

computer self-efficacy on the performance of the learning has been studied extensively in teachinglearning<br />

settings (Hill et al. 1986; Compeau et al., 1999; Lee, 2006). Vijayasarathy (2004) found that the<br />

computer self-efficacy have a positive effect on the behavioral intention to use systems as the more<br />

experience one gain online; the more important are concerns of control over personal information.<br />

Moreover, the individuals’ confidence in their computer-related knowledge and skills can influence<br />

perception on the ease or difficulty of carrying out a specific task using a new technology, and how useful<br />

that new technology will be (Vijayasarathy, 2004). Therefore, computer self-efficacy has been revealed<br />

by the literature to have a critical role in terms of its effect on perceive usefulness and perceived ease of<br />

use (Venkatesh et al., 1996; Hayashe, et al., 2003; Lee, 2006). Therefore the following hypothesis was<br />

developed:<br />

Hypothesis 6: Computer self efficacy has a positive effect on students and teacher’s perceived<br />

usefulness of educational technology in school at Kingdom of Bahrain.<br />

Hypothesis 7: Computer self efficacy has a positive effect on students and teacher’s perceived ease of<br />

use of educational technology in school at Kingdom of Bahrain.<br />

3.3.2 Motivation<br />

Motivation to perform a behavior can be divided into two main types: intrinsic and extrinsic motivation<br />

(Hennessy et al., 2005). Extrinsic motivation refers to drive of behaviors to achieve valued outcomes that<br />

are distinct from the activity itself (Deci and Ryan, 1985; Hennessy et al., 2005). Intrinsic motivation is the<br />

satisfaction gained from performing the behavior (Hennessy et al., 2005). According to Deci and Ryan<br />

(1985) perceived ease of use is a form of intrinsic motivation. Within the TAM, extrinsic motivation is<br />

capture by perceived usefulness, while intrinsic motivation related to the perceived ease of use.<br />

Motivation in the current study is measured by both intrinsic and extrinsic motivation, and then it is<br />

proposed to affect both perceived usefulness and perceived ease of use.<br />

Hypothesis 8: Motivation has a negative effect on students and teacher’s perceived usefulness of<br />

educational technology in school at Kingdom of Bahrain.<br />

Hypothesis 9: Motivation has a negative effect on students and teacher’s perceived ease of use of<br />

educational technology in school at Kingdom of Bahrain.<br />

3.3.3 Previous experience<br />

If a user has already using computer in prior job or education institutions, then it is believed that this<br />

experience will affect his/her perception toward using computer or any technology (Chang, 2008). The<br />

repetition of task improves task performance by reducing the effort required to perform the task (Alba,<br />

1987) and provides the user with a greater opportunity to consider different aspects of performing the<br />

behavior in a relatively objective manner (Kim, et al, 2009). Moreover, Dyck and Smither, (1996)<br />

revealed that as gaining more experience in using computer, user will become more computer<br />

confidence and will show a less negative attitude toward using computers (Trocchia and Janda, 2000;<br />

Dyck and Smither,1996). Thus, the previous experiences have a relationship with the intentional behavior<br />

toward using computer (Kim, et al, 2009) through the perceived usefulness and perceived ease of use.<br />

Moreover, a previous experience is proposed to be as a factor influence the computer anxiety.<br />

Hypothesis 10: Previous experience has a positive effect on students and teacher’s perceived usefulness<br />

of educational technology in school at Kingdom of Bahrain.<br />

Hypothesis 11: Previous experience has a positive effect on students and teacher’s perceived ease of<br />

use of educational technology in school at Kingdom of Bahrain.<br />

Hypothesis 12: Previous experience has a positive effect on student and teacher’s computer anxiety on<br />

using educational technology in school at Kingdom of Bahrain.<br />

12


3.3.4 Computer anxiety<br />

Jaflah Al-Ammary<br />

Computer anxiety is a specific anxiety because it is a feeling that associated with a person’s interaction<br />

with computers (Saade and Kra, 2006). Computer anxiety is defined as a state of mind of being fearful or<br />

nervous when using a computer. It includes factors such as who may have initiated the person to the<br />

computer technology (Brosnan 1998), success or failure experiences with hardware or software (Saade<br />

and Kira, 2006), and detract cognitive resources from task performance (Sam et al., 2005). Computer<br />

anxiety has been revealed by many studies to be affected by previous computer experience and affect<br />

ease of use of information technologies using the technology acceptance model (Bertrand and Bouchard,<br />

2008; Saade and Kira, 2006).<br />

Hypothesis 13: Computer anxiety has a negative effect on students and teacher’s perceived usefulness<br />

of educational technology in school at Kingdom of Bahrain.<br />

Hypothesis 14: Computer anxiety has a negative effect on students and teacher’s perceived ease of use<br />

of educational technology in school at Kingdom of Bahrain.<br />

4. Data collection and research variables<br />

The study sample is comprised of teachers and students from twenty six public schools (male and female<br />

schools) at Kingdom of Bahrain. These schools consist of primary, elementary, and secondary levels.<br />

Two hundred and four surveys were distributed to the twenty six schools. The survey instruments for this<br />

study were developed using validated items from the prior research as a means of assessing the<br />

theoretical constructs of the extended TAM model for educational technology, using the TAM scales of<br />

perceived usefulness, perceived ease of use, and behavioral intention from Davis et al. (1989) and Davis<br />

(1998). A scale for measuring subjective norms was developed using the measurement of Venkatesh et<br />

al. (2003), scales used for measuring computer anxiety and motivation were adopted from the<br />

measurement proposed by Saade and Kira (2006), and scale for measuring computer self-efficacy was<br />

developed using the measures of both Lee (2006) and Karsten et al. (1998). However, measurements of<br />

previous experiences were developed by the author for the purpose of the current research.<br />

5. Data analysis and results<br />

A Partial Least Squares (PLS) was applied to test the current model using Smart-PLS 2.0. PLS is a<br />

structured equation modeling method that analyzes how the items load on their constructs simultaneously<br />

with estimating all the paths in the model (Chiu et al., 2005). Data analysis in the current research<br />

possessed in two stages. In the first stage, the measurement model was evaluated to validate the<br />

reliability and validity of the constructs. In the second stage, the structure model was estimated using<br />

hypotheses testing to test the significance of the path coefficients.<br />

6. Demographics<br />

The survey instrument provides a response rate of 86% which can be considered as high rate bearing in<br />

mind the difficulty in getting the permission from ministry of education at Kingdom of Bahrain to be able to<br />

conduct the survey in the public schools. Demographic characteristics of the overall participants both<br />

teachers and students are presented in Table (1). The results in the table revealed that most of the<br />

participants either teachers or students are male (64.9%, 81% respectively) and are from secondary<br />

schools (57% and 56.5% respectively). Moreover, the results show that most of the teachers have B.Sc.<br />

and master degree (79.4% and 12.6%) and at least six years of experience in teaching (79.6%). Most<br />

importantly, it has been revealed from the results that most of the participated teachers have ICDL or<br />

other computer certificates (57.9%). In addition, the results show that most of the students have taken<br />

class using educational technologies (62.4%) which include the computer (50%), interactive board (30%)<br />

and smart board (20%).<br />

Table 1: Selected characteristics of the sample<br />

Teacher Percentage Students Percentage<br />

Gender Gender<br />

Male 64.9% Male 64.9%<br />

Female 35.1% Female 35.1%<br />

Level of School Level of School<br />

Primary 71.42% Elementary 71.43%<br />

Elementary 28.57% Secondary 28.57%<br />

Secondary Age<br />

13


Jaflah Al-Ammary<br />

Teacher Percentage Students Percentage<br />

Qualification 10-13 33.7%<br />

B.Sc. 79.4% 14-17 45.5%<br />

Post-diploma 9.0% >17 20.8%<br />

Master 12.6%<br />

PhD<br />

ICDL or any certificate in<br />

0.0% Take class using educational<br />

technologies<br />

Computer<br />

Yes 62.4%<br />

Yes 57.9% No 37.6%<br />

No 42.1%<br />

Take class using educational<br />

technologies<br />

Year of experience Computer 50%<br />

1-5 20.4% Smart board 20%<br />

6-10 43% Interactive board 30%<br />

>10 36.6% others 5%<br />

7. Assessing the measurement model<br />

The strength of the measurement model is determined by its reliability and validity. Cronbach’ alpha was<br />

used to assess the reliability value of each dimension as demonstrated in Table (2). All the reliability<br />

values are higher than 0.7, except that of subjective norms (0.636). However, these values are accepted<br />

because they are closed to 0.7. Convergent validity was assessed by the examination of composite<br />

reliability and Average Variance Extracted (AVE) (Chin et al., 1995). The data indicates that the<br />

measures are robust in term of their internal consistency reliability. The composite reliabilities of the<br />

different measures ranged from 0.744 to 0.886 which exceed the recommended threshold value of 0.7 for<br />

each construct. Moreover, results in Table (3) show that AVE values were all above the recommended<br />

range (0.50) (Chin et al., 1995), with values ranging from 0.524 to 0.668, thereby establishing convergent<br />

validity for each construct. Moreover, to assess the convergent validity confirmatory, factor analysis with<br />

Varimax rotation was conducted to assess the underlying structure for the items of each research<br />

construct. The loading of each factor should be greater than or equal to 0.5 which has been achieved<br />

(results not shown in the current paper).<br />

Table 2: Reliability, and convergent validity<br />

Items Cronbachs Alpha Composite Reliability AVE<br />

Behavioral Intention 0.819 0.882 0.653<br />

Subjective Norms 0.636 0.744 0.524<br />

Perceived Usefulness 0.775 0.859 0.671<br />

Perceived Ease of use 0.761 0.857 0.668<br />

Computer Anxiety 0.779 0.853 0.596<br />

Previous Experience 0.774 0.842 0.579<br />

Computer self efficacy 0.762 0.832 0.553<br />

Motivation 0.838 0.882 0.653<br />

8. Structural model<br />

The causal relationships in the proposal research model were tested. Consistent with Chin et al., (1995)<br />

bootstrapping was applied to produce standard error and t-statistics. This permits us to measure the<br />

statistical significance of the path coefficients. The statistical objective of PLS is to show high R and<br />

significant t-values, thus rejecting the null hypothesis of no effect. The t-values need to be significant to<br />

support the hypothesized paths. R indicates the explanatory power of the latent endogenous variables.<br />

Explanation of variance for each equation in the hypothesized model is presented in Table (3), while<br />

properties of the causal paths, including standardized path coefficients and t-values are presented in<br />

Table (4).<br />

14


Table 3: Explanation of variance<br />

Jaflah Al-Ammary<br />

Factor R<br />

BI 0.511<br />

PEOU 0.447<br />

PU 0.485<br />

CA 0.210<br />

As expected by the TAM literature, perceived usefulness, perceived ease of use and subjective norms<br />

have a great impact on the behavioral intention to accept and use educational technology. Three of them<br />

together accounted for a high variance of behavior intention to use educational technology (R=0.511).<br />

Moreover, perceived ease of use affected perceived usefulness. Hence H1, H2, H3, and H4 which are<br />

based on TAM (except hypothesis related to subjective norms) were all supported (0.200 (2.071), 0.455<br />

(2.90), 0.18 (1.47) and 0.246 (2.903), respectively). On the other hand, perceived usefulness was directly<br />

influenced by the computer self efficacy, motivation, and computer anxiety. Thus, H6, H8, and H13 are<br />

supported. However, subjective norms and previous experience has no effect on the perceive usefulness.<br />

Thus H5 and H10 are not supported. Furthermore, the results show that perceived ease of use was<br />

directly influenced by computer self efficacy and previous experience, while motivation has no effect of<br />

this factor. Therefore, H7, H11, and H13 were supported while H9 was rejected. Finally, the results show<br />

that there is a negative impact for previous experience on computer anxiety.<br />

Table 4: Model testing results<br />

Hypothesis Path coefficient T-value<br />

H1 Perceived usefulness - Behavior Intention 0.200 2.071<br />

H2 Perceived ease of use - Perceived usefulness 0.455 5.05<br />

H3 Perceived usefulness - Behavior Intention 0.184 1.472<br />

H4 Subjective Norms - Behavior Intention 0.246 2.903<br />

H5 Subjective Norms - Perceived usefulness 0.064 0.623<br />

H6 Computer self efficacy Perceived usefulness 0.203 1.499<br />

H7 Computer self efficacy Perceived ease of use 0.239 1.570<br />

H8 Motivation Perceived usefulness 0.389 2.95<br />

H9 Motivation Perceived ease of use 0.069 0.525<br />

H10 Previous experience Perceived usefulness 0.091 0.681<br />

H11 Previous experience Perceived ease of use 0.222 1.881<br />

H12 Previous experience Computer anxiety -0.459 5.691<br />

H13 Computer anxiety Perceived usefulness -0.203 1.499<br />

H14 Computer anxiety Perceived ease of use -0.239 1.570<br />

9. Discussion and conclusion<br />

The main conclusion of the current research is that factors which were hypothesized to have an impact<br />

on the acceptance and use of educational technology in schools are fallen within two categories:<br />

behavioral and Information Technology factors. As such behavioral factors include: motivation, subjective<br />

norms and computer anxiety while Information Technology factors include computer self efficacy, and<br />

previous experience. In the first place, the results revealed that teachers and students behavioral<br />

intention to accept and use educational technology is influenced indirectly by the proposed behavioral<br />

and information technology factors via perceived usefulness and perceived ease of use. Moreover, the<br />

results show that perceived usefulness and perceived ease of use have a great positive effect on both<br />

teachers and student’s behavioral intention to accept and use the educational technology. In addition, it<br />

has been revealed by the results that computer self efficacy and computer anxiety are the only factors<br />

that have a positive indirect effect on the behavioral intention to accept educational technology through<br />

15


Jaflah Al-Ammary<br />

both perceived usefulness and perceived ease of use. Conversely, motivation has a positive indirect<br />

effect on the behavioral intention through perceived usefulness only, while previous experience has<br />

indirect effect on behavioral intention through perceived ease of use. Subjective norms, on the other<br />

hand, have been identified by the research to be an important predictor for the behavioral intention to use<br />

educational technology as it shown a great direct effect on the behavioral intention. However, most of the<br />

literature and especially those that have been conducted in the context of Kingdom of Bahrain (Al-<br />

Ammary 2008, 2010a, b) revealed that the subjective norms have an impact on the perceived usefulness<br />

of IT, the findings of the current research does not support that. It has been revealed from the results that<br />

teachers and students will not feel that the educational technology can simplify their work and make it<br />

more effective if their peers or important people are using or think to use it. Finally, the results show that<br />

even by having a good experience in using computer or any educational technology, teachers and<br />

students will still express a fear of being under surveillance and believe that their wok conditions will<br />

worsen with the introduction of such technologies.<br />

Schools of Future project which aim at establishing a fundamental change in the way teaching and<br />

learning takes place, and transferring the traditional classroom into an open, interactive environment<br />

based on a wide-ranging use of technology, is the first step toward the educational reform at Kingdom of<br />

Bahrain. Hence, the research results can present a good insight of the wider range of issues that<br />

academic institutions would need to address when adopting educational technology in schools. Schools<br />

and ministry of education leaders at Kingdom of Bahrain should invest in a strategic plan for the adoption<br />

of educational technology which should be focused more on developing the human resources and<br />

building IT skills and capabilities within the school. Actually the ministry of education has provided a great<br />

effort to enhance the technical skills of the teachers in all schools and especially those that involved in<br />

the KHSFP. For instant, ministry of education has integrated ICT in education program to be accessible<br />

to the teachers and students to gain computer skills. Moreover, ICT training programs have been<br />

developed with the aim to empower number of schools to move in pioneering on new ways of learning.<br />

The programs include number of teacher training courses on digital literacy including the use of key tools<br />

such as Microsoft office, visual basic and web-based applications. These programs will help teachers to<br />

get a certified qualification and to use their ICT ability in their daily classroom life.<br />

Acknowledgement<br />

I would like to thank my senior students including: Bayan Abdulla, Mona Ali, and Nawal Isa for the effort<br />

that they have provided in collecting the data used for the current research.<br />

References<br />

Ajzen I. and Fishbein, M., (1980), Understanding Attitudes and Predicting Social Behavior, Prentice Hall, Englewood<br />

Cliffs, New Jersey, 1980<br />

Ajzen, I., (1991), “The Theory of Planned Behavior”, Organizational Behavior and Human Decision Processes, Vol.<br />

50, pp. 179-211<br />

Al-Ammary, Jaflah (2010a),”Factors affecting accepting and using of computer and internet by elderly people at<br />

Kingdom of Bahrain”, in proceeding of the International <strong>Conference</strong> of Management and Evaluation (ICIME<br />

2010), Cape Town, South Africa, 25-26 March<br />

Al-Ammary, Jaflah and Hamad, Sherifa (2008), “Factors influencing adoption of E-learning at University of Bahrain”,<br />

in proceeding of the International <strong>Conference</strong> on Information Technology (ACIT 2008), University of Sfax,<br />

Tunisia, 16-18 December<br />

Al-Ammary, Jaflah, Al Rowaili, Fatima, Al Noaimi, Latifa, Dodhy, Mona, and Zain, Maasoma (2010b),”Building<br />

Professional Learning Community in the secondary schools: a step toward the educational reform at Kingdom<br />

of Bahrain”, in proceeding of the 7 th International <strong>Conference</strong> on Knowledge Management (ICKM 2010),<br />

Pittsburg, Pennsylvania, 22-23 October<br />

Alba, J.W. and Hutchinson, J.W. (1987), “Dimensions of consumer expertise”, Journal of consumer research, Vol. 13,<br />

Issue 3, pp. 411-454<br />

Bertrand, Manon and Bouchard, Stephane (2008), “Applying the technology acceptance model to VR with people<br />

who are favorable to ITS use”, Journal of Cyber Therapy and Rehabilitation, Vol. 1, Issue 2<br />

Brosnan, Mark (1998), ‘The impact of computer anxiety and self efficacy upon computer performance’, The Journal<br />

of Computer Aided learning, Vol.14, pp. 223-234<br />

Chang, Su-Chao and Tung, Feng-Chang (2008), “An empirical investigation of students’ behavioral intentions to use<br />

the online learning course website”, British journal of education technology, Vol. 39, Issue 1<br />

Chin, C. and Todd, P. (1995), “On the use, usefulness and ease of use of structural equation modeling in MIS<br />

research: a note of caution”, MIS quarterly, Vol. 19, Issue 2<br />

Chiu, C., Hsu, M., Sun, S., Lin, T., and Sun, P., (2005), “Usability, quality, value, and E-learning continuance<br />

decisions.”, Computers and Education, Vol. 45, Issue 4, pp. 399-416<br />

Compeau, D.R. and Higgins, C.A.,(1995), “Computer self-efficacy: development of a measure and initial test”, MIS<br />

Quarterly, Vol. 19, Issue 2, pp.189-211<br />

16


Jaflah Al-Ammary<br />

Compeau, D.R. and Huff, S., (1999), “Social cognitive theory and individual reactions to computing technology: a<br />

longitudinal study”, MIS Quarterly, Vol. 23, Issue 2, pp.145-158<br />

Davis, F.,(1998), “Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology”,<br />

MIS Quarterly, Vol. 13, Issue 3, pp. 319-339<br />

Davis, F.D., Bagozzi, R.P, and Warshaw, P.R ,(1989),”User acceptance of computer technology: a comparison of<br />

two theoretical models”, Management science, Vol. 35, pages 982-1002, 1989.<br />

Doll, W.J, A. Hendrickson and X. Deng (1998), “Using Davis’s Perceived Usefulness and Ease-of-use Instruments for<br />

Decision Making: A Confirmatory and Multi-group Invariance Analysis”, Decision Sciences, Vol. 29, Issue 4, pp.<br />

839-869<br />

Dyck, J. L. and Smither, J. A (1996), “Older adults' acquisition of word processing: the contribution of cognitive<br />

abilities and computer anxiety”, Computer in Human Behavior, Vol. 12, Issue 1<br />

ESCWA (2005), National Profile for the information society in Bahrain, Economic and Social commission for Western<br />

Asia, United Nation, report, available online on http://www.escwa.un.org/wsis/docs/bahrain-2005-e.pdf<br />

Fishbein, M. and Ajzen, I.,(1975), Belief, attitude, intentions and behavior: an introduction to theory and research,<br />

Addison-Wesley, Reading, MA<br />

Havelka, D., (2003), “Students Beliefs and Attitudes Toward Information Technology”, In The Proceedings of<br />

ISECON 2003, Vol. 20 (San Diego), Revised in Information Systems Education Journal, Vol. 1, Issue 40, 2003.<br />

Hayashe, A.; Chen, C.; Ryan, T. and Wu, J., (2004),”Role of Social Presence and Moderating Role of Computer Self<br />

Efficacy in Predicting the Continuance Usage of E-Learning Systems”, The Journal of Information System<br />

Education, 2004<br />

Hennessey, B.A. and Amabile, T.M. (2005), “Extrinsic and intrinsic motivation”, in Nicholson, O.Audia, and M. Pillutia<br />

(Eds), Blackwell Encyclopedia of Mnagement, second edition Malden, MA., Blackwell publishing<br />

Hill, T.; Smith, N.D. and Mann, M.F., (1986),“Communicating Innovations: convincing computer phobic’s to adopt<br />

innovative technologies”, Advances in Consumer Research, Vol. 13<br />

ICE47 (2001), Main challenges facing the educational system at the beginning of the 21st century, UNESCO report,<br />

available online at: www.ibe.unisco.org/international/ICE47/English/natreps/Bahrain-en-part-2.pdf<br />

Karsten, R. and Roth, rm. (1998), “Computer self efficacy: a practical indicator of student computer competency in<br />

introducing IS courses’, Information Science, Vol. 1 Issue 3. Pp. 61-68<br />

Khan, Ted and Taber Ullah, Linda (1997),” learning by design: integrating technology into the curriculum through<br />

student multimedia design project”, New Horizons for Learning, Quarterly journal, available online at<br />

HTTP:www.newhorizons.org<br />

Kim, Young; Chun, Jae and Song, Jaeki (2009), “Investigating the role of attitude in technology acceptance from an<br />

attitude strength perspective”, International Journal of Information management, Vol. 29<br />

Lee, Y. C., (2006),”An empirical investigation into factors influencing the adoption of e-learning system”, Online<br />

Information review, Vol.30, No. 5<br />

Saade, R. and Kira, D. (2006), “The emotional state of technology acceptance”, Issues in Information Science and<br />

Information Technology, Vol. 3, pp. 529-540<br />

Sam, H.K.; Othman, A.E. and Nordin, Z.S. (2005), “Computer self-efficacy, computer anxiety, and attitudes toward<br />

the internet: a study among undergraduates in Unimas”, Educational Technology and Society, Vol. 8, pp. 205-<br />

209<br />

Segars, A., and Grover, V. (1993), “Re-examining perceived ease of use and usefulness: A confirmatory factor<br />

analysis”, MIS Quarterly, Vol. 17, Issue 4, pp. 517-525.<br />

Shaker, A. (2008), “Empowering Bahrain’s teachers and students to succeed in the digital future”, Partners in<br />

learning Middle East and Africa, available online<br />

Srite, Mark (2006), “Culture as an explanation of technology acceptance differences: an empirical investigation of<br />

Chinese an US users”, Australasian Journal of Information Systems, Vol. 14, Issue 1<br />

Taylor, S. and Todd, P.A., (1995 a),“Understanding information technology usage: a test of competing models”,<br />

Information Systems Research, Vol. 6 Issue 2, pp. 144-76<br />

Taylor, S. and Todd, P.A., (1995 b),“Assessing IT usage: the role of prior experience”, MIS Quarterly, Vol. 19, Issue<br />

4, pp. 561-70<br />

Trocchia, P. and Janda, S. (2000), “A phenomenological Investigations of Internet usage among older individuals”,<br />

Journal of Consumer marketing, Vol. 17, Issue 7, pp. 605-616<br />

Venkastesh, V. and Morris, M.G. (2000), “Why don’t men ever to stop to ask for directions? Gender, social influence,<br />

and their role in technology acceptance and usage behavior”, MIS Quarterly, Vol. 24, Issue 1<br />

Venkatesh, V. and Davis, F.,(2000),”A Theoretical Extension of the Technology Acceptance Model: Four Longitudinal<br />

Field Studies”, Management Science, Vol. 46, Issue 2, pp. 186-204<br />

Venkatesh, V. and Davis, F.D., (1996), “A model of the antecedents of perceived ease of use: development and test”,<br />

Decision Sciences, Vol. 27, pp. 451-481<br />

Venkatesh, V., Morris, M. G., Davis G. B., and Davis, F. D.,(2003), “User Acceptance of Information Technology:<br />

Toward a Unified View”, MIS Quarterly, Vol. 27, Issue 3, pp. 425-478<br />

Vijayasarathy, L. R., (1996), “Predicting consumer intentions to use on-line shopping: The case for an augmented<br />

technology acceptance model”, Information and Management, Vol. 41, Issue 6, 747-762, 2004.<br />

Wu, I. L. and Chen, J. L.,(2005), “An extension of Trust and TAM model with TPB in the initial adoption of on-line tax:<br />

An empirical study”, International Journal Human-Computer Studies, Vol. 62, 2005.<br />

17


Comparison of Feature Selection Techniques Using Fully-<br />

Controlled Simulation-Based Datasets<br />

Suzan Arslanturk 1 , Mohammad Siadat 1 , Theophilus Ogunyemi 1 , Ishwar Sethi 1 and<br />

Ananias Diokno 2<br />

1<br />

Oakland University, USA<br />

2<br />

Beaumont Hospital, USA<br />

sarslant@oakland.edu<br />

siadat@oakland.edu<br />

ogunyemi@oakland.edu<br />

isethi@oakland.edu<br />

adiokno@beaumont.edu<br />

Abstract: Data mining is the discipline of systematically reviewing datasets to determine what patterns,<br />

concurrences and/or rule sets can be discovered. In an effort to understand the effectiveness of a proposed mining<br />

methodology a well defined dataset is required for model verification. To this end, a fully-controlled simulation was<br />

created by which several different feature selection algorithms were evaluated. In this paper, we present a<br />

comprehensive comparison between attribute selection methods when noise, missing values and multicollinearity are<br />

in question. Our results show that "Relief" and "information gain" have outperformed other feature selection methods<br />

available in Weka when considering both sensitivity and specificity measures. We have evaluated the following<br />

features selection methods: J48, Relief, information gain, consistency based feature selection and correlation based<br />

feature selection to see which one handles additive noise better. The sensitivity of consistency based feature<br />

selection was 11% higher than the average sensitivity of other methods. However, it's specificity was 37% lower than<br />

that of the average. It is important to note that sensitivity or specificity alone does not give enough support to a<br />

method to say that it is the best way to handle the data. The best method, when both sensitivity and specificity are<br />

considered, was information gain. This method outperformed the average of other methods by 1% and 20.5% when<br />

we considered its sensitivity and specificity, respectively. Also in this regard, a goal of our study was to see which<br />

feature selection method outperforms within the missing set of values. In this case, when looking again at sensitivity<br />

and specificities, Relief and information gain proved to outperform the other methods of our study by 7.2% and<br />

12.4%, respectively. Our studies also show that when multicollinearity is embedded into the fully controlled dataset<br />

without any noise and missing values, the correlation based feature selection outperforms other methods. In<br />

summary Relief and information gain performed the best in all three situations in terms of its sensitivity and<br />

specificities.<br />

Keywords: attribute selection, rule extraction, classification, comparison techniques, simulation<br />

1. Introduction<br />

Feature selection is a machine learning method that selects an optimal subset of features by eliminating<br />

the ones which contain less predictive information. Reducing the dimensionality of a feature space<br />

improves the performance by diminishing the curse of dimensionality effect. Feature selection also gains<br />

advantage from efficiency in terms of storage and computational costs. Also, the execution time spent for<br />

both training and testing phases will decrease.<br />

The feature selection methods are categorized into three different forms: filter, wrapper or embedded<br />

(Molina 2002). It is important to evaluate the existing methods and figure out which one performs better in<br />

certain situations. Some algorithms perform well on correlated data while others can handle noise or<br />

missing values depending on the nature of the data. In filtering methods, feature selection is a<br />

preprocessing step independent of the induction algorithm. Information gain algorithm determines the<br />

importance of each feature by evaluating the uncertainty reduction, while ReliefF is sampling an instance<br />

and evaluating the difference between its nearest neighbors from both the same and the opposite class.<br />

Relevance scores are assigned to each attribute. Correlation based attribute selection methods eliminate<br />

one of the less important attribute that correlates with another attribute (Hall 1998). In embedded<br />

methods, inducer has its own feature selection algorithm embedded such as J48, a widely used decision<br />

tree algorithm. The occurrence of an attribute in a tree provides information about the importance of that<br />

particular attribute. Information gain and entropy reduction methods can be applied to each candidate<br />

feature of the decision tree node to evaluate the importance of each attribute (Sugumaran 2006). There<br />

have been several studies on feature selection methods. Hall and Holmes (Hall 1998) use the UCI<br />

dataset, a real world data set, that contains different data types such as categorical, continuous and<br />

multivariate. Six attribute selection methods are compared in terms of classification accuracy, reduction<br />

rate and speed. Molina et al., (Molina 2002) use a simulation based randomly created binary and nominal<br />

18


Suzan Arslanturk et al.<br />

valued dataset. Different types of syntactic functions are applied to the dataset to generate the class<br />

labels. Since it is a fully controlled scenario varying number of relevant, irrelevant and redundant features<br />

are placed in the dataset. However, the simulated dataset doesn’t lend itself to real longitudinal dataset of<br />

clinical trials and the functions used in this paper does not care about any particular attribute when<br />

deciding about the resultant, which does not fit to what we usually see in real world biomedical datasets.<br />

In this research, on the other hand, several feature selection algorithms are applied to a simulation based<br />

dataset with longitudinal trials in order to compare the performances of the algorithms in terms of different<br />

noise levels and different missing value levels and combination of both. Multicollinearity is added to the<br />

dataset to evaluate the most robust algorithms when dependencies between attributes are in question.<br />

Same attribute selection techniques are also applied to the combination of the longitudinal datasets. A<br />

simulation dataset is chosen since full control over the dataset is achieved.<br />

This paper is organized as follows. In Section 2 we explain different attribute selection techniques.<br />

Section 3 outlines how the database is created and how the rules are embedded into the dataset. In<br />

Section 4 the programming language used to embed the rules and the data mining tool to perform the<br />

feature selection algorithms are explained. Section 6 and Section 7 present the results.<br />

2. Attribute selection techniques<br />

Attribute selection techniques can be categorized into three different categories: filter, wrapper and<br />

embedded methods. In filtering methods, the feature selection method takes place before any learning<br />

algorithm. The undesirable attributes are filtered out before the classification step. All the training data is<br />

used in filtering methods (Hall 2003). In embedded methods, the learning algorithm has its own feature<br />

selection algorithm embedded in it (Molina 2002). J48 decision tree classification algorithm is a common<br />

example of an embedded method. In wrapper mode, on the other hand, the feature selection algorithm<br />

uses the learning algorithm as a sub-routine (John 1994).<br />

Five different attribute selection methods are applied to the dataset. Wrapper methods which are<br />

correlation based feature selection and Relief, a filtering method, information gain and an embedded<br />

method J48 decision tree based feature selection are applied to the dataset and the results are compared<br />

in terms of sensitivity and specificities.<br />

2.1 Correlation based feature selection<br />

Correlation based feature selection evaluates the dependencies between attributes and eliminate the<br />

ones which are correlated to each other. The irrelevant and redundant data has to be removed as much<br />

as possible. After the feature selection, the remaining data has to be highly correlated with the class and<br />

uncorrelated with each other.<br />

As equation 1 (Ghiselli, 1964) formalizes :<br />

Merits = (1)<br />

The feature subset S contains k different features where rfc is the feature to class correlation and rff is the<br />

feature to feature correlation. In order to have a good feature selection algorithm the merit has to be<br />

maximized.<br />

Symmetrical uncertainty can be evaluated as follows where H(X) and H(Y) are marginal entropies.<br />

Symmetrical uncertainty = 2.0 x (2)<br />

2.2 Consistency based feature selection<br />

The consistency of the class is evaluated by first figuring out all different combinations of the attributes.<br />

For each different combination the consistency is calculated by differentiating the number of occurrences<br />

of a particular attribute from the cardinality of the majority class (Hall 1998).<br />

19


2.3 Information gain<br />

Suzan Arslanturk et al.<br />

Information gain feature selection method the uncertainty of the class is evaluated with and without the<br />

attribute observation (Hall 1998).<br />

H(C) = - (3)<br />

H(C|A) = - (4)<br />

2.4 Relief<br />

Relief algorithm assigns a relevancy score to all the attributes in descending order. The algorithm selects<br />

an instance in each iteration and finds the nearest neighbor from the same and opposite class. The<br />

results should differentiate if the neighbor and the instance selected are in opposite classes (Hall 1998).<br />

The algorithm can handle noise if the neighbor number, k, is increased.<br />

2.5 J48 decision tree based feature selection<br />

J48 decision tree is a classification method that can also be used for feature selection. Each node of a<br />

tree involves an attribute and the occurrence of each attribute provide information about the importance<br />

of that particular attribute (Sugumaran 2006).<br />

The importance of each attribute can be evaluated by applying the information gain formula to each node<br />

of the tree.<br />

3. Dataset<br />

10 different simulation datasets are created each containing 20 attributes and 1500 subjects. In order to<br />

gain full control over the dataset the values in the dataset are formed to create rules (Agraval 1994). The<br />

last column of the datasets specify the resultants i.e. the class labels. Resultant is defined as the<br />

continence, expressed as 0, or incontinence, expressed as 1, of the subject. Resultant values are stored<br />

in the column space of the matrix.<br />

Subjects are defined as an individual and are represented in the row space of the matrix. The number of<br />

subjects is a function of the number of rules and the partition size for each rule. For example with ten<br />

rules and two hundred subject partition size there would be two thousand subjects represented in the<br />

matrix, with subjects one to two hundred being in the first partition, subjects two hundred and one to four<br />

hundred in the second, etc.<br />

Having defined the base dataset, it is now possible to manipulate the dataset for further methodology<br />

testing. There are two types of modifications that can be performed to the data set for this testing. First is<br />

the inclusion of noise to the data set. The second type of modification is to incorporate missing values,<br />

which shall be defined as an attribute missing a value.<br />

The expected results of the feature selection algorithms are the attributes that are the entities of the rules<br />

embedded to the datasets.<br />

A simulation dataset is used to allow maximum flexibility in creating and manipulating the data sets. The<br />

simulation allows for one base data set to be processed with multiple noise levels, multiple missing value<br />

levels and a mixture of both noise and missing values. The advantage of this methodology allows the<br />

researcher to understand the impact of varying levels of these factors on feature selection or any other<br />

metric of interest. Further, we can determine up to what missing value and noise levels a given data<br />

mining tool would be reliable. This is extremely important when one needs to use the tool on a real<br />

application. The current version of the matrix creation algorithm was intentionally restricted to binary data<br />

only. This was chosen primarily to validate the feature selection methods that were the driving force<br />

behind creating this simulated data. With the methodology described above modification to alternative<br />

data types, such as categorical and/or continuous, will be possible.<br />

4. Experimental methodology<br />

Our experiment applied the five different feature selection algorithms to the simulation datasets that were<br />

created using MATLAB. To apply the attribute selection methods Weka (Waikato Environment for<br />

20


Suzan Arslanturk et al.<br />

Knowledge Analysis) is used. Weka is an open source Java based data mining tool that performs pre<br />

processing, classification, clustering and many other machine learning tools by using a graphical user<br />

interface. First the dataset is uploaded to the Weka Explorer mode and the data is discretized. Then all<br />

the attribute selection algorithms are performed. Sensitivity and specificity are used to evaluate the<br />

performances in a statistical manner. Sensitivity i.e. true positive rate; measures the proportion of actual<br />

positives which are correctly identified. While specificity (1 – false positive probability) measures the<br />

proportion of negatives which are correctly identified. The algorithms with higher sensitivity and higher<br />

specificity are the ones that performs significantly better than the others.<br />

5. Results and discussions<br />

Our experiment applied five different feature selection algorithms to the simulation dataset and the results<br />

are compared when noise, missing values and multicollinearity are added to the data. The first set of<br />

experiments was designed to evaluate the following feature selection methods to see which one handles<br />

additive noise better: J48, Relief, information gain, consistency based feature selection and correlation<br />

based feature selection.<br />

In order to refer to a feature selection algorithm as robust, it has to both has a high sensitivity and a high<br />

specificity. Fig. 1 shows that J48 algorithm performs well in terms of sensitivity but there is a huge<br />

decrease in the specificity curve (Fig. 2) which makes the algorithm less desirable than the others.<br />

Without any noise Information gain and ReliefF algorithms both perform well. Consistency based feature<br />

evaluation has a low specificity and J48 decision tree classification algorithm has a low sensitivity. The<br />

results do not change in Cfs, Information Gain and Relief algorithms when the noise level is 2%, 5%,<br />

10%, 15%, ,respectively. In spite, in J48 decision tree the specificity decreases and the sensitivity<br />

increases rapidly when the noise level increase. The sensitivity of consistency based feature selection<br />

was 11% higher than the average sensitivity of other methods. However, its specificity was 37% lower<br />

than that of the average. It is important to note that data method that maximizes both the sensitivity and<br />

specificity is of interest. The best method, when both sensitivity and specificity are considered, was<br />

information gain. This method outperformed the average performance of the other methods by 1% and<br />

20.5% when we considered its sensitivity and specificity, respectively. Consistency based subset<br />

evaluation and J48 algorithms can not handle noise. The next set of experiments was designed to<br />

evaluate the feature selection methods to see which one handles missing values better. Fig. 3 and Fig. 4<br />

show the results of different missing value levels. In<br />

this case, when considering sensitivity and specificity, Relief and information gain were proved to perform<br />

better compared to the other methods of our study by 7.2% and 12.4%, respectively. Despite, the<br />

Consistency based feature selection and J48 algorithms can not handle missing values effectively. Our<br />

studies also showed that when multicollinearity was embedded into the dataset without any noise and<br />

missing values, the correlation based feature selection method outperformed other methods. In summary,<br />

Relief and information gain were the best in all three situations (noise, missing value, multicollinearity)<br />

when both sensitivity and specificity were considered.<br />

Figure 1: Noise-sensitivity<br />

21


Figure 2: Noise - specificity<br />

Figure 3: Missing values-sensitivity<br />

Figure 4: Missing values-specificity<br />

Suzan Arslanturk et al.<br />

22


6. Conclusions and future work<br />

Suzan Arslanturk et al.<br />

This paper presented a comparison between different attribute selection techniques by using a data<br />

mining tool, Weka. The results show that using different attribute selection techniques help the system by<br />

reducing the curse of dimensionality, time and storage. There is no single attribute selection technique<br />

that gives the best results. The advantages of each technique differs in different situations. One might<br />

give the most accurate results in a certain situation but the storage or time to run the algorithm might take<br />

too long which will effect the efficiency of the algorithm. However, the results show that the Relief and<br />

information gain algorithms give the most accurate results when missing values and noise are added.<br />

This study shows that if in a given dataset noise level is beyond 15% or missing value is beyond 12% the<br />

selected attributes are not reliable even when one uses Relief or information gain algorithms.<br />

The simulation was developed to allow for a particular randomly created matrix to be manipulated with<br />

differing noise, missing value, multicollinearity and combinations of both. Further development of this<br />

simulation model building tool will need to include categorical and continuous data, to allow for more<br />

realistic simulation models.<br />

References<br />

Agrawal, R., Srikant, R. (1994) Fast Algorithms for Mining Association Rules, Proceedings of the 20 th VLDB<br />

<strong>Conference</strong> Santiago, Chile.<br />

Ghiselli, E. E. (1964) Theory of Psychological Measurement.<br />

Hall, M., Holmes, G. (1998) Benchmarking Attribute Selection Techniques For Discrete Class Data Mining,<br />

Transaction on Knowledge and Data Engineering, Vol. 15, No. 3.<br />

Hall, M., Smith, L.A. (2003) Feature Selection for Machine Learning: Comparing a Correlation Based Filter Approach<br />

to the Wrapper, American Association of Artificial Intelligence.<br />

John, G., Kohavi, R. and Pfleger, K. (1994) Irrelevant features and the subset selection problem, in Proceedings of<br />

the International <strong>Conference</strong> on Machine Learning, San Francisco, CA.<br />

Molina, L., Belanche L., Nebot A.(2002) Feature Selection Algorithms: A survey and Experimental Evaluation.<br />

Second IEEE International <strong>Conference</strong> on Data Mining<br />

Sugumaran, V., Muralidharan, V., Ramachandran, K. I.(2006) Feature Selection Using Decision Tree and<br />

Classification through Proximal Support Vector Machine for fault Diagnostics of Roller Bearing, Mechanical<br />

Systems and Signal Processing.<br />

23


A Delphi-Multi-Criteria Decision Making Approach in the<br />

Selection of an Enterprise-Wide Integration Strategy<br />

Tridip Bardhan, James Ngeru, and Richard Pitts Jr<br />

Morgan State University, Baltimore, USA<br />

tridip.bardhan@morgan.edu<br />

jmngeru@gmail.com<br />

richard.pitts@morgan.edu<br />

Abstract: Driven by the need to be competitively effective and agile, the last few decades have witnessed explosive<br />

interest in the field of enterprise integration (EI). In spite of a great deal of research and advances in the EI discipline,<br />

there still seems to lack a comprehensive decision making framework which can assist decision makers during the<br />

crucial task of selecting an appropriate integration strategy that would optimally address most of the organization’s<br />

business and technical integration needs. Consequently, this paper attempts to fill the research gap by formulating<br />

and illustrating an application of such a framework through a case study. Since selection of an EI strategy can be<br />

regarded as a complex multi-criteria decision making problem that involves multiple participants who must weigh in a<br />

set of business and technical factors to evaluate different EI strategies, an integrated multi-criteria decision approach<br />

which combines Delphi techniques and the Analytic Network Process (ANP) is adopted. The use of the ANP is<br />

preferred over other multi-criteria decision analysis methods due to its ease of implementation and it provides a<br />

structure that allows for an intensive analysis of the criteria which shows the effects of various criterions to be taken<br />

into consideration. The Delphi technique on the other hand helps to ensure a consensus in the decision making<br />

process by allowing active participation of various decisions makers and assimilation of their different schools of<br />

thought. The formulated framework is then illustrated as a case study of a state government agency that is<br />

embarking on enterprise-wide integration. The choice of the case study in the public sector setting was preferred<br />

since these organizations are generally divided into several organizational functions; hence, such strategic decisions<br />

usually involve different decision makers who represent their functional units. As a result, EI decisions which are<br />

generally made at an enterprise-wide level are more challenging in public organizations. The results from the case<br />

study indicate that there exists a conflict of interest among decision makers who participate in the EI strategy<br />

evaluation and selection process. However, with the application of the proposed model, an effective decision making<br />

process can be realized with significant consensus among decision makers. The use of the proposed model in this<br />

context is a ground-breaking concept and represents a scientific and systematic approach to decision making that<br />

could minimize risks and improve the benefits generally associated with the selection of an enterprise-wide<br />

integration strategy.<br />

Keywords: enterprise integration, multi-criteria analysis, delphi technique, analytic network process<br />

1. Introduction<br />

Today’s business environment is characterized with globalization, customization, digitalization,<br />

virtualization and agility (Huang and Fan, 2007). Organizations are competing and creating new<br />

opportunities in global markets with global partners and consumers. The competitive and dynamic nature<br />

of the new business playground means new technologies methods and strategies are needed. To be<br />

competitive, businesses in both private and public sector must have enough agility to respond to<br />

business needs and with effective service delivery. This on the other hand is only possible with rapid<br />

sharing of information among business processes both within and across organization. For the last few<br />

decades, Enterprise Integration (EI) has emerged as a revolutionary concept and solution that promises<br />

to assist organization with achieving this goal. EI aims to link separate business processes giving them<br />

increased leverage (Lam and Shankararaman, 2004). Thus it is not surprising that for the past few years,<br />

survey after survey of top CIO have ranked EI initiatives, such as Service Oriented Architecture (SOA)<br />

adoption, Enterprise Resource Planning (ERP) adoption/upgrades, Data integration, in their list of top ten<br />

priorities (Gartner-inc, 2009; Nascio, 2009). To many top executives, EI is now a business necessity.<br />

However, implementation of an EI strategy is not an easy task. Past researchers have called for thorough<br />

evaluation of organization’s requirements before undertaking EI implementation (Themistocleous, 2004).<br />

Thus, this paper aims at presenting formulation and application of a systematic methodology for analysis<br />

in the selection of an EI strategy. Selection of an EI strategy is a multifaceted problem involving more<br />

than the choosing a vendor or integration software system. It is important to consider the organization’s<br />

vision, requirements and constraints; the technical merits and limitations of the chosen strategy.<br />

In spite of remarkable interest in overall EI concept, little has been done in development of a framework<br />

for EI implementation, especially in EI strategy selection. Most of the EI research has been centered in<br />

advancing the technology to enable integration. Janssen argues that organizational issues and<br />

stakeholders interests should be addressed in the selection process (Janssen and Cresswell,<br />

24


Tridip Bardhan et al.<br />

2005).Thus, the need to develop a comprehensive framework incorporating a compendium of<br />

organizational factors during the selection of an EI strategy is indispensable (Giachetiti, 2004).<br />

The methodology discussed in this paper utilizes multi-criteria analysis principles, specifically integrating<br />

ANP method with Delphi technique. Analytic Network Process (ANP) developed by Saaty (1980) was<br />

chosen for its ability to incorporate multiple factors with different relative weights and with dependencies<br />

characteristics. This gives the decision makers the ability to prioritize the importance of one factor over<br />

another. ANP has been used in complex problems involving technology selection, and thus it is not a far<br />

to see that it may be applicable in this situation. To further improve the results on using ANP within this<br />

context, Delphi technique is applied. The Delphi technique improves the process by facilitating active<br />

participation and consensus among evaluation process partakers representing different organization’s<br />

functions.<br />

The paper proceeds with an overview on EI, followed by methodology and a case study.<br />

2. An overview of Enterprise Integration (EI)<br />

Frye and Gulledge (2007) termed EI as “alignment of strategies, business processes, information<br />

systems, technologies, and data across organizational boundaries to provide competitive advantage”.<br />

Lack of this alignment is source of many organizational problems. The process of achieving integration<br />

involves all managerial and technological factors that enable cross-functional process integration (Frye<br />

and Gulledge, 2007). Integration is approached in various ways and at various levels, such as, physical,<br />

application, business and can also be achieved through enterprise modeling approach or any<br />

methodological approach facilitating consistent enterprise-wide decision making (Chen et al., 2008). EI is<br />

also classified on whether it involves integrating processes and applications within or outside<br />

organizational boundaries. Inter-organizational enterprise integration, attempts to integrate business<br />

processes between enterprises (B2B), such as Supply Chain Management systems (SCMs), or electronic<br />

purchasing processes (e-procurement). Intra-organizational integration is the integration of applications<br />

within the organization, attempting to integrate custom applications and packaged systems (Silveira and<br />

Pastor, 2006)<br />

Lammer et al (2008) acknowledged existence of a number of integration approaches and concepts,<br />

differentiated by integration’s level and architecture. However, Spackman et al (2004) and Chandra et al<br />

(2009) argued that beyond the marketing hype; “there is a great deal of overlap between the integration<br />

approaches”. Thus, at highest level of integration, while devising integration strategy for the enterprise,<br />

there are limited numbers of overall approaches available (Spackman and Speaker, 2004). EI<br />

approaches can therefore be broadly narrowed down into four basic categories, including (Spackman and<br />

Speaker, 2004; Chandra and Juarez, 2009): (i) Point-to-point based integration; (ii) ERP integration; (iii)<br />

Enterprise Application Integration (EAI) approach; (iv) and Service oriented integration approach.<br />

In early stages of integration project, determination of strategic integration approach is among the most<br />

important and challenging planning task. Decisions made during this process impacts negatively or<br />

positively most of the subsequent development activities, and consequently overall success and quality of<br />

the resulting system. This decision activity is usually complicated by the fact that it involves different<br />

participants, contributing in decision making process and who have different goals, criteria and perception<br />

about certain alternatives. The activity is further challenging as it comes in early stages of project<br />

lifecycle, hence hard to reason about the consequences of the decision made (Al-Naeem et al., 2005).<br />

2.1 Factors Influencing adoption of particular integration approach<br />

EI projects are strategic investments, and should be closely linked to organization vision, goals and<br />

strategies. These projects involve high expenditure, risk and with great impact in almost every<br />

organization’s aspects. Various organizations develop and implement integration projects motivated by<br />

completely different integration drivers. The initial rationale for EI adoption influences problem definition,<br />

methods of achieving goals and other subsequent activities (Wei et al., 2005).<br />

Lam et al (2007) categorized integration drivers into two categories namely; organizational drivers and<br />

project drivers. Puschmann et al (2001) identified five integration drivers including; software, financial,<br />

internal and external drivers. Themistocleous (2001) argued that many organizations are driven to<br />

integration initiatives based on organizational, operational, technical and strategic perceived benefits.<br />

These drivers and benefits should be taken into consideration when an enterprise is devising and<br />

25


Tridip Bardhan et al.<br />

evaluating different integration strategies (Themistocleous and Irani, 2001; Janssen and Cresswell, 2005;<br />

Themistocleous, 2004).<br />

As EI involves an attempt to connect a number of different types of software products, the selection of<br />

integration approach should not only be based on functional and business requirements, but should also<br />

take into consideration non-functional and other technical requirements. Non-functional requirements<br />

describe properties that a system should possess. They are defined by stakeholders and greatly<br />

influence adoption of a particular integration approach. Non-functional requirements should not be<br />

overlooked, since doing so could lead to a poor system quality, unsatisfied stakeholders and eventually<br />

unsuccessful investment (Al-Naeem et al., 2005; Al-Naeem et al., 2004). Spackman et al (2004)<br />

proposed a Software Quality Attribute Trading (SQUAT) technique, which takes into consideration known<br />

software qualities attributes to evaluate integration approaches and technologies. Silveira et al (2006)<br />

developed EAI evaluation tool by further characterizing and describing international software quality<br />

evaluation standard (ISO-9126) for integration tools. Some of these quality attributes includes;<br />

adaptability, platform Neutrality, scalability, security, reliability, modifiability, performance, interoperability,<br />

maintainability, flexibility and testability<br />

In general, selection and evaluation of integration approaches should take into consideration multiple<br />

factors, which include business, organizational, project and technical related factors (Bardhan and Ngeru,<br />

2009), as shown in Figure 1.<br />

Figure 1: EI approach factors (Bardhan and Ngeru, 2009)<br />

3. Methodology<br />

Evaluation and selection of integration approach is a complex and challenging process. It involves<br />

consideration of several organizational factors, from both technical and business perspective. Moreover,<br />

the process consists of several contributors representing different levels of organizational functions and<br />

with varying goals and preferences. Thus, the problem requires an elaborate and systematic approach<br />

that would allow participants involved in the process to evaluate different alternatives by considering and<br />

expressing their preferences on multiple factors. That been the case then, application of multi-criteria<br />

decision analysis (MCDA) techniques is warranted.<br />

A plethora of MCDA methods exists, where most common includes; Analytic Hierarchy Process (AHP),<br />

Analytic Network Process (ANP), Simple Additive Weight (SAW), Simple Multi-Attribute Rating Technique<br />

(SMART), Elimination and Choice Translating Reality (ELECTRE) and Technique for Order Preference<br />

by Similarity to Ideal Solution (TOPSIS) (Triantaphyllou, 2000). Guitouni (2004) argued that the purpose<br />

of every MCDA method is to assist in making a good recommendation. However, not all MCDA methods<br />

produce a good recommendation for all situations; since every method’s theoretical and axiomatic<br />

development is based on some assumptions and hypothesis. Given the nature of EI approach selection<br />

problem, ANP is deemed as the most appropriate method to utilize during the evaluation process. ANP is<br />

preferred for this case due to its ability to provide an elaborate structure that allows for an intensive<br />

analysis on problem elements and the resolution of inter-dependency characteristics among the<br />

elements. However, ANP framework does not fully support group decision making process. Since the EI<br />

approach selection process involves a group of contributors with varying views and preferences, a sense<br />

of consensus among the contributors is an important element in ensuring the process overall success. It<br />

is therefore posited that incorporation of Delphi technique in ANP framework, illustrated in Figure 2, will<br />

facilitate vigorous involvement of process contributors and bring forth a sense of unanimity among them.<br />

26


Tridip Bardhan et al.<br />

More discussion about ANP can be found in (Saaty, 1980; Saaty, 1996) and on Delphi technique in<br />

(Brown, 1968; Hartman et al., 2007; Okoli and Pawlowski, 2004).<br />

Figure 2: Delphi-ANP evaluation methodology<br />

The Delphi-ANP methodology presented in this article consists of four Delphi rounds, which are blended<br />

in with ANP procedure to yield the following eight steps; (i) Problem formulation and factor identification;<br />

(ii) Forming evaluation panel; (iii) Defining evaluation Criteria; (iv) Structuring the problem network model;<br />

(v) problem elements appraisal; (vi) judgment weights analysis; (vii) elements ranking approval; (viii) and<br />

finally the best alternative selection (See Figure 2)<br />

To illustrate the ANP-Delphi methodology, this paper will consider an actual case study in a state<br />

government agency, referred hereinafter as TGA (real name withheld).<br />

4. Application of Delphi-ANP methodology for EI approach selection<br />

The Delphi-ANP methodology was applied and validated at TGA, a state government agency in eastcoast<br />

of USA. TGA is composed of several organizational functions and manages several engineering<br />

and transportation facilities across the state. However, many of these functions are supported by different<br />

systems for their day-to-day operations. Unfortunately, most of these systems exist within technical and<br />

functional walls of separation that creates barriers to integration and sharing of data; and thus inefficient<br />

business processing. Accordingly, the agency has taken several measures to address its integration<br />

challenges as an effort to achieve its strategic goals of effectiveness, efficiency and service improvement.<br />

At the time of this study, TGA was embarked in efforts to integrate some of its backbone information<br />

systems, such as those used in agency’s financial management, supply chain management and<br />

maintenance operations. In other similar integration projects, TGA has successfully developed in-house<br />

integration solutions by either establishing direct interfaces between applications, replicating data<br />

between databases or simply databases-consolidation. However, for the particular initiative considered in<br />

this study, any of these traditional approaches may not have been viable option for several reasons. First,<br />

the systems intended to be integrated are the backbone of the agency’s operations; and must therefore<br />

integrate with other systems in a different platform and located within and outside the agency’s boundary.<br />

Secondly, these systems uses different data format, and thus the issue of data uniformity presented a<br />

27


Tridip Bardhan et al.<br />

major integration challenge. Finally, these systems are in constant upgrade to the latest versions, which<br />

also requires any direct interface existing between them to be upgraded or redeveloped, which is rather<br />

laborious and expensive.<br />

Consequently, the management was seeking for a forward-looking approach that will allow future<br />

expansion of the integrated network system, and resonate with organization’s strategic goals. The<br />

approach must also provide sound, cost effective and long-term solution that is efficient and meets all of<br />

the stipulated needs of the agency. As a start, some high level integration approaches were suggested.<br />

These suggested approaches consisted; (1) Implement a Service oriented based integration, (2) Adopt a<br />

vendor-based EAI adapter, (3) Implement a single enterprise system such as ERP that would either<br />

replace these systems or integrate them and (4) built in-house integration solution that establishes direct<br />

interface links between the systems.<br />

Deciding on the approach to commit the SGA presented a great challenge to the management, as these<br />

approaches were perceived with varying strengths and weakness. The formulated Delphi-ANP<br />

methodology was therefore used in assisting the TGA management to decide the appropriate EI<br />

approach for their agency.<br />

The Delphi-ANP methodology process involved the following steps;<br />

Step 1: Problem formulation: The step comprised of preparatory and necessary tasks to help<br />

understanding the problem domain and set forth the evaluation process. The initial task is to set the<br />

objectives and goals of the evaluation process. Then, a review of literature, as well as interviews with<br />

domain experts was conducted in order to broadly identify feasible integration approaches and a list of<br />

criteria, which will possibly be used in gauging the identified alternatives.<br />

Step 2: Evaluation panel establishment - This step involved setting up a team of participants to<br />

contribute in the evaluation process. In a Delphi evaluation process, a group of five to twenty willing<br />

experts (Rowe and Wright, 2001; Armstrong, 2001) is generally considered adequate. Thus, a panel of<br />

six members consisting of two IT planning manager, two project managers and two IT analysts was<br />

established. The members selected were believed to be knowledgeable about the integration subject and<br />

organizational requirements.<br />

Step 3: Selection of organization’s specific criteria factors: - In this step, also marked as Delphi first<br />

round, the evaluation panel, formed in the previous step, was presented with an overview explaining the<br />

problem at hand, the purpose of the evaluation task, the expectations and procedures of the entire<br />

evaluation task process, and a list containing description of feasible alternatives and possible evaluation<br />

criteria. The panel was requested to select and justify factors that they would consider important in<br />

evaluating the presented alternatives. They were also requested for additional factors that may have<br />

been excluded from the list. From this round, the group identified a list of eighteen factors, categorized as<br />

strategic related, cost related, technical related and other organizational factors. The final list of the<br />

approved criteria is presented in Table 1.<br />

Step 4: Structuring problem into an ANP network model: In this step, the problem was configured to<br />

form initial ANP network structure, which consists of all the problem elements (i.e. the goal, criteria and<br />

alternatives) and also indicating any relationship existing between the elements.<br />

This step also entailed Delphi second round, where the evaluation panel was presented with proposed<br />

ANP network structure as well as Delphi first round results, which contained other members’ opinions<br />

(anonymous feedback) and any of their discrepancies. The members were given an opportunity to<br />

reconsider their judgments’. The results from this round were similar to the result from round one, and<br />

thus, consensus on evaluation criteria, presented in<br />

Table 1 and final problem structure formed, shown in Figure 3.<br />

Step 5: Appraisal of problem’s model elements: - First, a questionnaire was developed to assess the<br />

relationships between the elements of the network structure, through pair-wise comparison. The<br />

questionnaire was instituted in MS-excel application so as to facilitate analysis, like checking consistency<br />

on the fly (see Figure 4).<br />

28


Table 1: Evaluation criteria for EI approaches<br />

Tridip Bardhan et al.<br />

Step five was also marked as Delphi round three, where the participants were presented with instructions<br />

on how to use excel-ANP template and the final problem network structure developed in the previous<br />

step. Using Saaty’s 1-9 scale (Saaty, 1980), the participants were requested to gauge (i) relative<br />

importance of criterions with respect to goal; (ii) relative preference of alternatives with respect to each<br />

criterion and; (iii) any relative influences existing between criterions within and outside the cluster.<br />

29


Tridip Bardhan et al.<br />

Figure 3: Integration approach selection problem network system<br />

Figure 4: Pair-wise questionnaire built in MS Excel<br />

30


Tridip Bardhan et al.<br />

Step 6: Gathered judgment weights analysis: - The pairwise comparison was done using MS-excel<br />

template, with in-built capability to calculate elements local priorities and pairwise-matrix consistency.<br />

Thus, the first task of the analysis step involved obtaining and organizing each panel member’s pairwise<br />

comparison matrices, and their respective element priority vectors. The pairwise matrices were also<br />

assessed for the completeness and consistency (see Figure 5).<br />

Figure 5: Pair-wise comparison matrix as built in Microsoft Excel ®<br />

The element priorities were calculated using Saaty’s (1980) additive normalization algorithm, which<br />

consists of the following three steps:<br />

For each pairwise comparison matrix :<br />

Where is pairwise matrix from panel member , for cluster and containing relative measures,<br />

assigned to element when compared to element (Note that and )<br />

Sum the values of each column in comparison matrix i.e.<br />

Divide each element in matrix with respective column sum i.e.<br />

pairwise comparison matrix.<br />

to obtain normalized<br />

Sum each row’s element and divide row’s sum by elements in the row, to<br />

obtain vector (Eigen vector) containing estimates of relative priorities of the elements in the<br />

cluster i.e.<br />

Where: is the local priority for element of cluster , for panel member<br />

The consistency of each pairwise comparison matrix was checked using Saaty’s (1980) consistency<br />

index (CI) calculated as:<br />

31


Where<br />

Tridip Bardhan et al.<br />

RI = consistency random index of randomly generated reciprocal matrix form scale 1-9, (see table 2), and<br />

Where,<br />

And<br />

Aggregating members’ priorities: - Since the local element priorities varied from one panel member to<br />

another, there was a need to aggregate the results for all the participants. Saaty (1980) argued that<br />

geometric mean can be used to correctly represent a group of experts’ consensus. Thus, geometric<br />

mean, calculated using below equation, was used to aggregate participants’ element priority into a group<br />

element priority.<br />

Where: = the group priority of element in cluster<br />

Building super-matrix – To resolve interdependencies characteristics existing between the elements of<br />

the problem network system and eventual derivation of final elements’ priorities, ‘supermatrix’ concept<br />

proposed by Saaty (1996) is applied. A supermatrix is partitioned matrix, where each supermatrix’s block<br />

is composed of a set of relationships between elements as represented by the problem model. The<br />

elements’ relative weights obtained in the preceding steps, which indicates relative importance and<br />

influence of elements with each other, are entered in their respective block of the supermatrix, to form unweighted<br />

supermatrix, shown in Figure 7. The un-weighted supermatrix’s columns are not stochastic i.e.<br />

each column of supermatrix does not add to one. Thus, to obtain stochastic supermatrix (shown in Figure<br />

8), each block of un-weighted supermatrix is multiplied by corresponding cluster priority (for this study<br />

case, all the clusters had relative-equal weights). To obtain long-term stable set of weights, the stochastic<br />

supermatrix is successively raised with large powers, usually 2k+1 (where k is a large arbitrarily number),<br />

until the process converges to a limit supermatrix. The final priority weights of the elements are obtained<br />

by normalizing the limit supermatrix, shown in Figure 9.<br />

Table 2: Average random consistency (Saaty, 1980)<br />

32


Tridip Bardhan et al.<br />

Figure 6: Combining of individual element priority using geometric mean<br />

Figure 7: Un-weighted super-matrix<br />

Step 7: Element Ranking Approval – In this step, marked as fourth and last Delphi round, a report<br />

describing the analysis results was prepared and sent to the evaluation participants for their review. The<br />

report contained all the elements ranking based on the final priority weights obtained from normalized<br />

limit supermatrix. The report underscored the elements’ ranking, by indicating each alternative<br />

performance with respect to the criterions and each criterion relative importance with respect to the goal.<br />

Any major discrepancies between members’ weights was highlighted, and members were given a last<br />

opportunity to revise and justify their pairwise relative judgments’, especially where notable discrepancy<br />

exists. In this last Delphi round, two members revised some of their pairwise comparison, arguing that,<br />

while undergoing through the evaluation process, they have become more acquainted on EI related<br />

issues, and therefore felt the need to make adjustment to reflect their true judgments. The final element<br />

rankings are presented in Table 3, Figure 10, Figure 11 and Figure 12.<br />

33


Figure 8: Stochastic (weighted) super-matrix<br />

Tridip Bardhan et al.<br />

Figure 9: Normalized limit super-matrix<br />

Step 8: Selection of best alternative – This step marked the end of the evaluation process. The final<br />

report, which included members’ comments and changes requested from the previous step, was<br />

produced. The final report emphasized on the alternative ranking, see Figure 10, and more in particular to<br />

the first ranked alternative, which is actually the selected alternative.<br />

5. Results and discussion<br />

The use of Delphi-AHP evaluation approach provided a systematic framework in selection and evaluation<br />

of integration approaches. From the case study, EAI was ranked as the most appropriate approach to<br />

commit TGA (See Figure 10). The Delphi-ANP analysis presented members with comprehensive<br />

breakdown on how they reached to the final decision, by prioritizing factors with respect to goal, in Figure<br />

34


Tridip Bardhan et al.<br />

11, and by how each alternative relatively support different factors, as shown by the radar chart in Figure<br />

12.<br />

Table 3: Alternative performance with respect to criteria elements<br />

Figure 10: Alternative final score and ranking<br />

The case study results also showed that implementation cost was the major factor influencing the final<br />

decision, scoring 12.8%, followed by implementation time with 12%. Scalability and security were<br />

considered as the major technical factors with each scoring 7.8% and 7.6% respectively. In strategic<br />

perspective, business agility, EI approach overall alignment with organization strategy and cooperation<br />

with other business partners are leading factors, each scoring 6.8%, 5.3%and 5.3%, respectively.<br />

Overall, factors deemed to enhance organization interoperability were well favored; although ranked<br />

behind the factors related to approach implementation and security. These results are presented in Table<br />

3, Figure 10 and Figure 11.<br />

EAI approach scored average in almost every criterion considered. SOA approach was viewed as the<br />

most outstanding approach as far as organization’s strategy is concerned, but performed relatively poor<br />

in terms of cost related factors. ERP scored relatively well in technical related factors, especially on<br />

issues related to security, maintainability and complexity. However, it was viewed as an expensive<br />

approach. Finally, in-house application integration was favored in terms of cost related factors, although<br />

viewed as an expensive approach to maintain. In addition, in-house application integration ranked quite<br />

high in terms of organizational related factors such as acceptability and technical support/expertise<br />

availability.<br />

35


Tridip Bardhan et al.<br />

Figure 11: Criteria ranked by their percentage contribution to overall decision<br />

Figure 12: Radar chart showing alternatives performance with respect to criteria<br />

6. Conclusion and future work<br />

Decisions related to EI implementation and technologies are unstructured and complex, due to<br />

consideration of wide range factors, and involvement of different stakeholders. This paper demonstrated<br />

a Delphi-ANP evaluation approach, which systematically considers multiple factors and improves the<br />

ANP decision process by incorporating Delphi technique. The Delphi technique improved the process by<br />

first, allowing selection of requirements considered by the participants. Many factors included in the<br />

problem network mean more pair-wise comparison, and this may affect the process ability. The Delphi<br />

feedback also helped in improving consensus in the decision making process.<br />

The use of the Delphi-ANP approach allows EI selection to be logically and systematically structured<br />

within the context of the organization’s strategic vision. It enables decision makers to comprehensively<br />

consider the strengths and weakness of each EI strategy against various criteria and sub criteria.<br />

36


Tridip Bardhan et al.<br />

Although the process is quite demanding, its adoption allows productive and informative participation of<br />

different stakeholders with varying perspective to arrive at a consensus with proper documentation for<br />

such a complex decision.<br />

Additional testing of this model is encouraged in different environments. Additionally, sensitivity analysis<br />

to determine robustness of the decision should be performed.<br />

References<br />

Al-Naeem, T., Gorton I., Babar, M. A., Rabhi, F. and Benatallah, B. (2005) A Quality-Driven Systematic Approach for<br />

Architecting Distributed Software Applications.<br />

Al-Naeem, T., Rabhi, F. A., Benatallah, B. and Ray, P. K.. (2004) Systematic Approaches for Designing B2B<br />

Applications. International Journal of Electronic Commerce, 9, 41 - 70.<br />

Armstrong, J. S. (2001) Principles of forecasting: a handbook for researchers and practitioners, Springer.<br />

Bardhan, T. K. and Ngeru, J. (2009) “Application of Fuzzy Analytical Hierarchy Process in Enterprise Integration<br />

Approach Selection,” Proceedings (in CD-ROM) of the 14 th Annual International <strong>Conference</strong> on Industrial<br />

Engineering – Theory, Applications, and Practice, Anaheim, CA, October 18-21, 2009.<br />

Brown, B. B. (1968) Delphi Process: A Methodology Used for the Elicitation of opinions of Experts.<br />

Chandra, S. and Juarez, R. (2009) A Practical Approach to Enterprise Integration. Defense AT&L. Defence<br />

Acquisition University.<br />

Chen, D., Doumeingts, G. and Vernadat, F. (2008) Architectures for enterprise integration and interoperability: Past,<br />

present and future. Computers in Industry, 59, 647-659.<br />

Frye, D. W. and Gulledge, T. R. (2007) End-to-End Business Process Scenarios. Industrial Management \& Data<br />

Systems, 107, 749-761.<br />

Gartner-Inc (2009) Gartner EXP Worldwide Survey of More than 1,500 CIOs Shows IT Spending to Be Flat in 2009.<br />

Gartner Newsroom.<br />

Giachetiti, R. E. (2004) A Framework to Review the Information integration of Enterpise. International Journal of<br />

Production Research, 42, 1147-1166.<br />

Guitouni, A. (2004) Decision-aid to improve organisational performance. <strong>European</strong> Journal of Operational Research,<br />

162, 581-583<br />

Hartman, F. T., Krahn, J. and Skulmoski, G. J. (2007) The Delphi Method for Graduate Research. Journal of<br />

Information Technology Education, 6, 1-21.<br />

Huang, S. and Fan, Y. (2007). Model Driven and Service Oriented Enterprise Integration---The Method, Framework<br />

and Platform. In: ALPIT '07: Proceedings of the Sixth International <strong>Conference</strong> on Advanced Language<br />

Processing and Web Information Technology (ALPIT 2007), Washington, DC, USA. IEEE Computer Society,<br />

504--509.<br />

Janssen, M. and Cresswell, A. M. (2005) Enterprise Architecture Integration in E-government. In: 38th Hawaii<br />

International <strong>Conference</strong> on System Sciences.<br />

Lam, W. (2007) Enterprise Architecture and Integration: Methods, Implementation, and Technologies, Idea Group Inc<br />

(IGI).<br />

Lam, W. and Shankararaman, V. (2004) An enterprise integration methodology. IT Professional, 6, 40-48.<br />

Lammer, A., Eggert, S. and Gronau, N. (2008) A Procedure Model for a SOA - Based Integration of Enterprise<br />

Systems. International Journal of Enterprise Information Systems, 4, 1-12.<br />

NASCIO (2009) State CIO Priorities 2009. National Association of State Chief Information Officers Website.<br />

Okoli, C. and Pawlowski, S. D. (2004) The Delphi method as a research tool: an example, design considerations and<br />

applications. Information & Management, 42, 15-29.<br />

Puschmann, T. and Alt, R. (2001) Enterprise Application Integration - The Case of the Robert Bosch Group.<br />

Proceedings of the 34th Hawaii International <strong>Conference</strong> on System Sciences, 34.<br />

Rowe, G. & Wright, G. (2001) Expert Opinions in Forecasting. Role of the Delphi Technique. In: ARMSTRONG, J. S.<br />

(ed.) Principles of forecasting: a handbook for researchers and practitioners. Boston: Kluwer <strong>Academic</strong><br />

Publishers.<br />

Saaty, T. L. (1980) The Analytic Hierarchy Process, New York, NY, McGraw-Hill.<br />

Saaty, T. L. (1996) Decision Making with Dependence and Feedback: The Analytic Network Process, Pittsburgh,<br />

RWS Publications.<br />

Silveira, R. and Pastor, J. A. (2006) A Model For Enterprise Application Integration Tools Evaluation. In: <strong>European</strong><br />

and Mediterranean <strong>Conference</strong> on Information Systems (EMCIS), Costa Blanca, Alicante, Spain.<br />

Spackmanm, D. and Speaker, M. (2004) Enterprise Integration Solutions, Microsoft Press.<br />

Themistocleous, M. (2004) Justifying The Decisions For EAI Implementations: A Validated Proposition Of Influential<br />

Factors. Journal of Enterprise Information Management, 17, 85.<br />

Themistocleous, M. and Irani, Z. (2001) Benchmarking the Benefits and Barriers of Application Integration.<br />

Benchmarking: An International Journa, 8, 317 - 331.<br />

Triantaphyllou, E. (2000) Multi-Criteria Decision Making Methods: A comparative Study, Springer.<br />

Wei, C. C., Chien, C. F. and Wang, M. J. (2005) An AHP-Based Approach to ERP System Selection. Int. J.<br />

Production Economics, 96, 47-62.<br />

37


Selection, Implementation and Post Production of an ERP<br />

System<br />

Imran Batada 1 and Asmita Rahman 2<br />

1<br />

Institute of Business Administration, Karachi, Pakistan<br />

2<br />

University of Georgia, Georgia, USA<br />

Imran.batada@khi.iba.edu.pk<br />

Asmitab@uga.edu<br />

Abstract: Enterprise Resource Planning (ERP) is an enterprise level system that integrates all the business<br />

processes of an organization into a single database. ERP System enables all the departments of an organization to<br />

share and manipulate the data. ERP system has a centralized structure consisting of one single central database<br />

which can be accessed by all the department(s). This centralized structure is one of the major reasons of current<br />

ERP system‘s popularity. Selection of an ERP system plays a vital role in the implementation of ERP system. We<br />

used AHP method for the selection. Once proper selection has been made, the implementation process begins. This<br />

process entails; accurate requirement gathering, user training, technical training, testing, data migration etc. Our<br />

research will give an in-depth analysis and recommended approaches for the preparation of RFP document,<br />

selection of ERP system, selection of vendor, requirement gathering, system design, testing and final roll-out. The<br />

objective of our research is to ensure successful implementation of ERP system. Our research also proposes a<br />

potential model for implementing an ERP System. Our model proposes the organization to have two (2) major<br />

committees: (1) ERP Implementation Committee (2) ERP Steering Committee. ERP Implementation Committee will<br />

drive the complete implementation process. It must include members of all the department of the organization. Since,<br />

the system is for the functional users, it is essential that all the stakeholders are part of this committee as well. In<br />

addition, the ICT (Information & Communication Technology) department should also be a part of this committee.<br />

The other committee is the Steering Committee; this committee should consist of members of top management of the<br />

organization. All business process reengineering should be directed to the ERP Steering Committee. Steering<br />

committee is also responsible for supervising the overall progress of the implementation. They will also decide the<br />

level of customization necessary for the ERP System.<br />

Keywords: ERP Implementation process, selection of an ERP System, RFP development of an ERP system,<br />

proposed implementation model of an ERP system<br />

1. Introduction<br />

Enterprise Resource Planning System (ERP) is a centrally integrated system within all the departments of<br />

an organization. With the rapid evolution of web 2.0 and web 3.0, ERP has now become a necessity of<br />

today’s world. Due to its centralized nature and numerous benefits, more and more companies have<br />

adapted the ERP Systems. Implementation of an ERP system is a very significant and crucial process. It<br />

requires involvement of not only the implementation team, but also all the stake holders (including users<br />

etc). In this paper we will discuss the implementation process of an ERP System in detail and review the<br />

process with the help of a case study. We will also look at some of the common issues faced during the<br />

implementation with proposed solutions.<br />

2. Development of strategic and implementation committee<br />

Creation of a strategic and implementation committee plays a critical role in successful implementation of<br />

an ERP system. It is recommended to involve all the stake holders in the implementation process, as it<br />

not only gives them a better understanding of the system, but will also help modify the system throughout<br />

the process to meet the needs of the organization. It will also play a vital role during the post-production<br />

implementation process.<br />

3. Buy vs build<br />

While deciding on the type of the ERP system, there are two major options. An organization can either<br />

buy an existing ERP solution or they can build their own custom ERP system based on their<br />

requirements. This is one of the major decisions to be made by the organization based on their needs<br />

and available resources. (Kim J., 2009: 10) If an organization chooses to use an existing available<br />

solution, they must be a little flexible on the desired functionality. Even though, customization is an option<br />

while using an existing solution, too much customization is not recommended. (Zhao Y, and Fan S.,<br />

2007: 13). Surplus amount of customization may result in making the system unreliable. In addition,<br />

applying new patch and release may become a laborious and expensive operation.<br />

38


Imran Batada and Asmita Rahman<br />

On the hand if the organization chooses to build their own custom solution, they can be more obstinate<br />

about their requirements; however, building a new solution will take more time and such a system will<br />

also need time to mature. In addition, it usually costs more than using an existing available solution.<br />

Regardless, of whichever choice an organization makes, the decision should be made by the entire<br />

steering committee of the organization<br />

4. Development of Request for Proposal (RFP) document<br />

Development of the RFP document is important as this document states the basic requirements of the<br />

system. If the requirements are stated clearly and precisely, the implementation process will become<br />

easier. Therefore, it is important to spend considerable amount of time and effort for the development of a<br />

clear RFP document. A sample RFP document may look like the Figure 1.<br />

Figure 1: Popular RFP templates, 2010: 2<br />

The first column indicates the hierarchy of the desired modules. The second column indicates the<br />

criterion of the desired module. The third column indicates the priority of the desired module in the<br />

system (may range from 1-10, with 10 as the most important). The fourth column indicates if the given<br />

module is mandatory or optional. Mandatory implies that the organization must have the functionality in<br />

order to be able to proceed. These four (4) columns should be filled by the organization. The remaining<br />

six (6) columns should be filled by the vendor (in case of using an existing solution) or the developmental<br />

committee (in case of building a custom solution). Out of the box means that the functionality is already<br />

built-in into the System. Modification via supported means the functionality can be made available with<br />

minor configuration changes. Supported via third party solution means the functionality is not available in<br />

the system but it can be integrated with 3 rd party solution. Minor Customization means it can be achieved<br />

with minor changes in the coding. Major customization means it can be achieved with the major changes<br />

in the source code. FUT means the functionality is not available right now in the system but it will be<br />

available in the future release or future patch. NS means not supported in the system and it is also not<br />

recommended to do the changes in the source code for such functionality. It is basically core business<br />

processing changes.<br />

5. Selection of an ERP system<br />

The RFP document developed in the previous stage gives a clear picture of the desired features with<br />

their respective priorities. Once the requirements have been clearly stated, the implementation committee<br />

will need to select the ERP System. We can use methods like AHP method for selection of an ERP<br />

System (Kunz J., 2010: 7). Implementation committee must review each and every requirement stated in<br />

the RFP document and carefully propose to choose the vendor that satisfies most or all the desired<br />

39


Imran Batada and Asmita Rahman<br />

requirements (Lall and Teyarachakul, 2006: 1). While planning the level of customization, one must not<br />

plan the desired customization more than 10% (percent). Final decision about the selection should be<br />

made by steering committee based on the facts and the recommendations of the ERP implementation<br />

committee.<br />

5.1 Practical example<br />

The proposed method was successfully implemented in a higher education institute in Karachi, Pakistan.<br />

The university core business was Campus Solution System which they used to connect faculty and<br />

students. However, the existing system was not integrated and had multiple databases in it. In addition,<br />

proper reporting system was also lacking and duplicate efforts had to be made due to multiple databases.<br />

In order to resolve this issue, university decided to go for an ERP System. Management was also flexible<br />

to re-engineer their process and was willing to adopt best practices in the organization. (Quiscenti M,<br />

Bruccoleri M , La Commare U, Noto La Diega S, and Perrone G., 2006: 14).<br />

5.2 Selection Method for ERP system<br />

Steering committee created a team that included business users, functional owner and the IS team for<br />

selecting the system. Series of ERP System workshops were arranged for evaluation process. We<br />

selected core business functionality to evaluate the mandatory needs of the system for the organization.<br />

We selected AHP Method for selection an ERP System and incorporated the evaluation data in it. We<br />

identified nine major criteria and under each criterion relevant sub criteria that covered the complete<br />

expectation of the users (Shapira A, and Simcha M., 2009: 6). Figure 2 shows the results of three major<br />

systems that were considered.<br />

Figure 2: Results of three major systems<br />

40


Imran Batada and Asmita Rahman<br />

Figure 3 shows the priority values of each aspect. For example, functionality was assigned maximum<br />

weight-age as compared to other aspects. Questionnaires were prepared with a list of mandatory<br />

functionalities (Wei C, Chein C, and Wang M., 2005: 5).<br />

Figure 3: Priority values of each aspect<br />

The following figure (Figure 4) shows the final results based on the individual values from Figure 2 and<br />

Figure 3.<br />

Figure 4: Final results<br />

6. Selection of vendors<br />

While selecting the vendors based on the requirements, we may end up with more than one vendor who<br />

is providing a potentially suitable solution (Grossman T, and Walsh J. 2004:3). The following are the four<br />

major criteria to be considered while making the final decision:<br />

1. Must have similar business experience<br />

2. Must have recommendations by clients in the same geographical region<br />

3. Must have success stories of implementing same or similar solutions.<br />

4. Must have proven experience in providing ERP support service<br />

7. Recommended organogram for the organization<br />

Figure 5 shows the recommended organogram for the organization. Project director would be the core<br />

responsible person of the project; however ERP project manager will play a vital role both technically and<br />

functionally. He would also be responsible to conduct functional training session(s) for all the users.<br />

Figure 5: Recommended organogram<br />

41


8. ERP implementation process<br />

Imran Batada and Asmita Rahman<br />

A proposed ERP implementation process comprises of the following phases:<br />

8.1. Requirements Gathering Phase<br />

8.1.1 Requirements Workshop(s)<br />

8.1.2 Requirements Documentation<br />

8.1.3 Requirements Review and Signoff<br />

8.2. Technical Training Phase<br />

8.3. System Design Phase<br />

8.3.1 Solution Design Workshop<br />

8.3.2 Solution Design Documentation<br />

8.3.3 Solution Design Review and Signoff<br />

8.4. Testing Phase<br />

8.4.1 Training for the users<br />

8.4.2 Functional Training & Testing Phases<br />

8.4.3 Signoff of Testing Phase<br />

8.5. Data Migration<br />

Signoff from all the stake holders must be obtained before proceeding from one phase to the other.<br />

8.1 Requirement gathering phase<br />

8.1.1 Requirement workshops<br />

This phase involves requirement gathering. (Parr A, and Shanks G., 2000: 9) This is one the very<br />

important phases of ERP implementation process. All the functional head(s) should be a part of this<br />

phase. Inputs from all the stake holders must be taken. GAP Analysis will be identified during this phase<br />

as well. Organization should be flexible enough to change the business process in order to avoid too<br />

much customization in the system as excessive customization makes the system unreliable. Also, adding<br />

of new patches or updates may result in creating problem(s) for the organization. Roles and rights of the<br />

user(s) should also be decided during this phase.<br />

8.1.2 Requirement documentation<br />

In this phase all the requirements based on requirement workshop should be documented so that it can<br />

ensure that all the stake holders are on the same page. Since the GAP Analysis is also performed based<br />

on the requirement, stakeholders will clearly understand what they will get in the new system.<br />

8.1.3 Requirement review and signoff<br />

All the stake holders will review the document and signoff. It is very important to involve all the functional<br />

users to be part of this process, so that they can have a feeling of ownership of the system.<br />

8.2 Technical training<br />

This phase is very critical and it has to be performed for the ICT (Information and Communication<br />

Technology) Department. This phase will provide technical training to them which will enable them to give<br />

42


Imran Batada and Asmita Rahman<br />

their maximum input during the design and testing phases. The lower the dependency on the vendor, the<br />

more successful the system would be.<br />

8.3 System design phase<br />

8.3.1 Solution design workshop<br />

Screen shots or prototypes of the system should be created and showed to the functional users based on<br />

requirement documentation. The prototype would help the non-technical functional users to understand<br />

the functioning of the system as the snapshots would give a preview of the actual system to be<br />

implemented. It will also help them understand their input parameters for all the forms and reports. They<br />

will also see the validation of the forms. During this phase all the changes as per the design can be<br />

accommodated.<br />

8.3.2 Solution design documentation<br />

Based on the design workshop, a detailed documentation should be made for the system design<br />

workshop. All the changes should be accommodated as per the requirement of the functional owners.<br />

8.3.3 Solution design review and signoff<br />

Once all the changes in the document have been incorporated, all the functional heads should sign the<br />

document.<br />

8.4 Testing phase<br />

8.4.1 Training for the users<br />

Proper training sessions should be arranged in this phase so that the users can obtain a better<br />

understanding of the system. This will aid the testing of the system.<br />

8.4.2 Functional training and testing phase<br />

Testing phase is entirely dependent on the training of the system. If proper training sessions have been<br />

conducted not only for functional heads but also for functional user(s), they will be able to properly test<br />

the system based on the requirements and design phase. Users can test the system with test data during<br />

this phase. Proper test cases should be developed to thoroughly check the system during this phase. ICT<br />

Department should be capable to thoroughly check the system along with all the users.<br />

8.4.3 Signoff of testing phase<br />

Once the system has been thoroughly tested, all the functional users need to signoff.<br />

8.5 Data migration<br />

After successfully completing the testing phase, it is time to migrate data from the legacy system to the<br />

new System. Deadlines (Cut-off times) should be established. All the data should be validated before it is<br />

made live. It is recommended to run the legacy system in parallel for one financial year. This helps verify<br />

proper functioning of the new system.<br />

9. Final rollout<br />

In this phase the newly developed system is made live and all the users in the organization are directed<br />

to start using the new system. Again, it is recommended to keep the previous traditional system running<br />

in parallel for some time to ensure proper functioning. It can also serve as a back-up.<br />

10. Training<br />

Training phase is a continuous phase for all the organization’s users. (Scott J., 2006:11). The more<br />

involved users are in this phase, the more comfortable they will feel in using the new system. Resistance<br />

of the staff will gradually decrease as they learn to understand and use the new system.<br />

43


11. Post production of an ERP system<br />

Imran Batada and Asmita Rahman<br />

This is the crucial phase. It is important to have vendor’s technical support during this phase. A vendor<br />

resource available at the premises is a good, however, costly solution. Ideally, the in-house team should<br />

be able to handle the issues, in case vendor’s support is limited or not available. One must identify the<br />

post production cost before making the final decision as it may get very expensive. (Wei C, 2008:12). It is<br />

recommended to have at least one (1) year post production support in order to fully use the system to its<br />

fullest capabilities.<br />

12. Conclusion<br />

Selection of ERP is the major decision for the organization. It is a proven fact that a wrong decision in<br />

selection will adversely effect in the overall performance of the organization. Involvement of the<br />

stakeholders for selecting the product and using the right method for evaluating is the correct path for<br />

successful implementation of the ERP System. AHP Methodology is a recommended technique for<br />

complex decision making. AHP Methodology permits evaluator to evaluate priorities of alternative on the<br />

basis of performance of each condition (Teltumbdey A., 2000: 4).<br />

Selection of the right ERP System using the right methodology makes the whole implementation process<br />

easier. Implementation of ERP System in the university or in any organization who are in the business for<br />

more than 50 years is itself a tough task. It is recommended to re-engineer the process instead of doing<br />

major customization in the system.<br />

In this study nine major criteria and the sub criteria have been identified for evaluating the ERP System.<br />

University used AHP Model. First they developed the weight for the criteria, then they developed the<br />

rating for each decision alternative for each criteria and finally they have calculated the weighted average<br />

rating for each decision alternative (Laia V, Truebloodb R, and Wong B. , 1999: 8).<br />

Implementation of ERP not only involves money, but also requires time, effort and dedication from all the<br />

stake holders. If the stakeholders are not part of the decision making process then they will not feel the<br />

ownership of the new system. Therefore, it is highly recommended to involve them from the beginning.<br />

Training of ICT department is also very necessary, they should be technically sound as eventually they<br />

will be responsible to train the non-technical user(s).<br />

References<br />

Grossman T, and Walsh J. (2004) ‘Avoiding the pitgalls of ERP system implementation’, Information Systems<br />

Management<br />

Kim J. (2009) ‘Activity-based framework for cost savings through the implementation of an ERP system’, International<br />

Journal of Production Research<br />

Kunz J. (2010) 'The Analytic Hierarchy Process (AHP)', Eagle City Hall Location Options Task<br />

Laia V, Truebloodb R, and Wong B. (1999) 'Case Study: Software selection: a case study of the application of the<br />

analytical hierarchical process to the selection of a multimedia authoring system', Information & Management<br />

Lall V, and Teyarachakul, S. (2006) ‘Enterprise Resource Planning (ERP) System selection: A data envelopment<br />

analysis (DEA) Approach’, Journal of Computer Information Systems.<br />

Parr A, and Shanks G. (2000) ‘A model of ERP project implementation’, Journal of Information Technology<br />

Popular RFP Templates (2010), Discrete Enterprise Resource Planning (Discrete ERP) RFP Template, [Online]<br />

Available: http://rfp.technologyevaluation.com/store.asp [November 10, 2010]<br />

Quiscenti M, Bruccoleri M , La Commare U, Noto La Diega S, and Perrone G. (2006) ‘Business process-oriented<br />

design of Enterprise Resource Planning (ERP) systems for small and medium enterprises’, International Journal<br />

of Production Research<br />

Shapira A, and Simcha M. (2009) 'AHP-Based Weighting of Factors Affecting Safety on Construction Sites with<br />

Tower Cranes', Journal of Construction Engineering and Management<br />

Teltumbdey A.(2000) 'A framework for evaluating ERP projects', Int. J. Prod. Res.<br />

Wei C, Chein C, and Wang M. (2005) 'An AHP-based approach to ERP system selection', International Journal of<br />

Production Economics<br />

Scott J. (2006) ‘Post- Implementation Usability of ERP Training manuals: The user’s perspective’, Information<br />

Systems Management<br />

Wei C, (2008) ‘Evaluating the performance of an ERP system based on the knowledge of ERP implementation<br />

objectives’, Int J Adv Manuf Technology<br />

Zhao Y, and Fan S. (2007) ‘Implementation approach of ERP with mass customization’, International Journal of<br />

Computer Integrated Manufacturing<br />

44


Evolution of eProcurement Model in the Italian Public Sector:<br />

From Government Centralisation to Regional Delocalisation<br />

Clara Benevolo and Renata Paola Dameri<br />

University of Genoa, Italy<br />

benevolo@economia.unige.it<br />

dameri@economia.unige.it<br />

Abstract: In this work the authors present the results of a survey regarding public eProcurement in Italy. The Italian<br />

Government has been starting to drive the introduction of eProcurement practices in the public sector since the end<br />

of the twentieth century. Its aim was especially to rationalise the public spending by the benefits offered by<br />

eProcurement, that is: more competition between the suppliers, more transparency in carrying on the tenders and<br />

auctions, lower prices thanks to higher volumes, speeding up of purchasing process. To quickly reach these goals,<br />

the Government opted for a centralised solutions. At that time, it was the best one to support the implementation of a<br />

new technology and to reach synergies and scale economies. In a second and more mature phase, the Government<br />

decided to encourage a delocalisation of public eProcurement, especially involving the twenty Italian Regions to<br />

implement their own eProcurement platform. The aim was to pursue not only quantitative benefits, but also<br />

qualitative ones, especially regarding a better alignment between the amount and nature of public spending and the<br />

real needs of citizens. The second phase is in progress yet, but surveying the actual implementation of<br />

eProcurement all over the twenty Italian Regions some results are already evident. The authors use their research on<br />

the Italian case to explore potentials and benefits, barriers and difficulties of public eProcurement implementation in<br />

complex administrative organisations, such as the Italian State, with very high level of public spending. Empirical<br />

evidences and lessons learned are explored, to define a comprehensive interpretative model of eProcurement in the<br />

public sector.<br />

Keywords: eProcurement, public administration, decentralisation, business case study<br />

1. Introduction<br />

During the latest ten years, the diffusion of eProcurement has been increasing both in private and in<br />

public sector. Governments and local public agencies consider eProcurement as a powerful instrument,<br />

especially to control the public spending, to gain efficiency and save money. However, even if<br />

eProcurement is based on a quite simple technology, it is not so easy to implement, because it affects<br />

several aspects, such as: IT infrastructure, software implementation, business transactions design,<br />

process reengineering, people behaviours and the relationship between costs and benefits. Moreover,<br />

several public agencies implemented eProcurement without any methodological approach, without<br />

applying nor theoretical models nor best practices. For these reasons, both successes and failures can<br />

be obtained, but it is not clear which critical factors influence the reached results.<br />

In Italy, public eProcurement implementation can be divided in two phases.<br />

In the first one, the Government implemented eProcurement solutions, especially aiming to reduce public<br />

expenses and looking at granting affordable and secure transactions. EProcurement was a centralised<br />

practice, supported by the strong commitment by the State able to develop this innovative instrument. But<br />

the centralised practice forced local public agencies to apply the same model, not ever suitable with their<br />

specificity.<br />

In the second phase, eProcurement was shifted from central to local level, involving Regional and other<br />

local public agencies to implement their own eProcurement applications, choosing both technical<br />

instruments and organisational solutions. The main aim was to gain efficiency and in the meantime to<br />

create a better fit between eProcurement applications, organizational solutions and local needs.<br />

The Italian case is interesting for several aspects: first, because in Italy the public expense is very high<br />

and it is crucial to improve the control on public spending and the quality of purchasing; second, because<br />

the territorial organisation of public bodies in Italy is complex and structured on different levels; third,<br />

owing to the evolution of the general framework of eProcurement implementation.<br />

In this paper, we examine the evolution of the Italian public eProcurement model, from the central to the<br />

regional one, and the key lessons learned in both the implemented experiences. We identify both<br />

reached goals and implementation problems, both critical success factors and winning choices, in order<br />

to design an implementation framework able to be applied to similar public agencies in other Countries.<br />

45


2. Literature review<br />

Clara Benevolo and Renata Paola Dameri<br />

EProcurement is an important topic, emerging in the latest ten years, both in companies and in the public<br />

sector. It regards the use of Internet potential to redesign business process and to improve operations<br />

quality. The focus of eProcurement has ever been double:<br />

On one side, eProcurement is seen like an instrument to control purchases, to reduce cost, to save<br />

money;<br />

On the other side, eProcurement is seen like a driver for business change, process quality<br />

improvement and more effective management (Ash and Burn 2003; Dooley and Purchase 2006).<br />

Also in the public sector, eProcurement is considered an important weapon to reach several goals<br />

(Croom and Brandon-Jones 2005):<br />

A better control on public purchases and spending, cost reduction and more effective use of public<br />

resources;<br />

The improvement of public administrative processes, to reduce bureaucracy and to increase<br />

transparency in the management of public agencies;<br />

The increasing of the public programs effectiveness, better focusing choices and expenses on the<br />

more important needs for citizens.<br />

However, the eProcurement implementation in the public sector should face several obstacles:<br />

organizational and cultural barriers, such as difficulties in evaluating the eProcurement benefits.<br />

On the organizational aspect, to implement eProcurement practices requires not only to create<br />

effective and appropriate web sites (Bruno et al. 2005) nor to implement a technological platform to<br />

manage tenders, agreements and eMarketplace, but it requires to rethink the administrative activity in<br />

public administration, that is to redesign all the activities included in public procurement and<br />

introducing eProcurement solutions in a redefined process. Without the organizational redesign,<br />

eProcurement perhaps can produce economic benefits, but it is not possible to reach the other (and<br />

perhaps more important) goals: administrative process quality, transparency, suppliers competition,<br />

public choices effectiveness (Reddick 2004).<br />

On the cultural aspect, eProcurement needs not only ICT competences, both regarding the use of<br />

web sites and in data processing, but also to change the administrative practices and to accept to<br />

modify their own habits. On the contrary, suspicion and resistance on implementation are the<br />

stronger barriers to successful and profitable eProcurement implementation in the public sector<br />

(Vaidya, Sajeev and Calender 2006; Wirtz, Sebastian and Schierz 2010).<br />

On the evaluation aspect, eProcurement benefits are relatively easy to evaluate, if limited to cost<br />

savings, but more difficult to measure if regarding process quality, transparency, effectiveness of<br />

public actions (Singer et al. 2009). The lack of a well conceived framework to evaluate also the<br />

intangible and qualitative benefits of eProcurement not only prevents to understand the real impact of<br />

this business practice on the efficiency and effectiveness of public purchases and spending, but also<br />

delays the improvement of administrative processes to integrate eProcurement (Rendon 2008).<br />

Despite barriers and difficulties, eProcurement is spread in several public administrations, all over the<br />

world (Moon 2005). Also in Italy, eProcurement practices and solutions have been adopted since the end<br />

of XX century. During the latest ten years, eProcurement was driven along with two paths:<br />

To enlarge its adoption, increasing both the number of public agencies using eProcurement and the<br />

amount of transactions using this channel;<br />

To improve the quality of eProcurement, by better ICT solutions, people training, implementation<br />

frameworks able to integrate eProcurement in more effective administrative processes and<br />

operations (Polimi 2009).<br />

Also in Italy, eProcurement faced the same difficulties and barriers noted in the international environment<br />

(Carpineti, Piga and Zanza 2006; Somasundaram 2004). Better implementation frameworks and<br />

knowledge sharing on best practices in eProcurement implementation in the public sector could help to<br />

assess the present and potential benefits of this technology. Governments and local administrative<br />

agencies could improve gains deriving from their eProcurement solutions, realising better returns on<br />

these technological and organisational investments.<br />

46


Clara Benevolo and Renata Paola Dameri<br />

3. eProcurement in the Italian public sector: The Government centralisation<br />

Starting from the end of Nineties, the Italian Government has been starting an innovation process, aiming<br />

to introduce eProcurement in the public sector. The main goals of this innovation were to reduce<br />

expenses, to increase transparency of purchasing and to improve efficiency of administrative processes<br />

in the Public Administration. In this first phase of eProcurement implementation in the public sector, the<br />

role of the central Government was crucial: indeed, only the State was able to support this innovative<br />

practice, helping all the public agencies to introduce this new technology into their administrative<br />

activities.<br />

Moreover, the Italian Government grounded its own implementation strategy on centralisation, to reach<br />

both the best efficiency and the best effectiveness of the innovation process, minimising in the meantime<br />

the cost of the organizational and technological solutions. The Central Public Administration defined all<br />

the phases of purchasing process and the operations to be automated, and also the software applications<br />

to be used for this aim.<br />

During the first phase of its implementation, three aspects are important to understand the evolution of<br />

eProcurement in the Italian public sector:<br />

The need of laws and rules to govern this practice;<br />

The public body that drove the implementation activity;<br />

The methods and instruments used to implement eProcurement.<br />

EProcurement implementation in the pubic sector needed for the first to define laws and rules to be<br />

applied to public contracts, especially regarding:<br />

The award of contracts concluded on behalf of the State, regional or local authorities and other<br />

bodies governed by public law entities;<br />

The opening-up of public procurement to competition;<br />

The verification of the suitability of tenderers, in open procedures, and of candidates, in restricted and<br />

negotiated procedures;<br />

The electronic signature.<br />

The successfully implementation of eProcurement in public sector depends not only on electronic<br />

solutions, but also on the affordability and compliance of contracting procedures. In the first phase of<br />

public eProcurement implementation, the central Government played the crucial role of regulator, both<br />

issuing its own laws and implementing <strong>European</strong> Directives. At present, all the rules regarding electronic<br />

activities of Italian Public Administration are collected in the “Codice dell’Amministrazione Digitale (CAD)”,<br />

Digital Administration Code. It is a set of rules governing the use of Information Technology in all the<br />

relations between the Italian Public Administration and the citizens (D.Lgs. 07.03.2005, n° 82). In Europe,<br />

eProcurement is ruled by Directive 2004/18/EC of the <strong>European</strong> Parliament and of the Council of 31<br />

March 2004 on the coordination of procedures for the award of public works contracts, public supply<br />

contracts and public service contracts.<br />

By these means, the Italian State defined its own eProcurement reference model (Alsac 2007), including<br />

rules regarding: the administrative process to be applied to award a public contract by electronic<br />

technologies, the affordability of economic operators, the public contracts transparency and the digital<br />

signature (Figure 1).<br />

The reference model includes all the steps and activities needed to drive to the final signed contract. That<br />

is:<br />

1. The assessment of economic operators to be admitted to electronic tenders, auctions and markets,<br />

by checking their affordability and compliance to the law regarding commercial relations with public<br />

agencies;<br />

2. The general framework of laws and rules to be applied to carry out all the procurement phases,<br />

complying with all the qualifications of a public tender such as transparency, accountability and so on;<br />

3. The procedure to be followed for a regular award of a valid public contract;<br />

4. The digital signature, for validating the contract.<br />

47


Clara Benevolo and Renata Paola Dameri<br />

Figure 1: The Italian public eProcurement reference model<br />

To better support the launch of eProcurement and its first implementation, the Italian Government choose<br />

a centralised model and set up a public agency, called Consip, to concentrate all the activities regarding<br />

public eProcurement. Consip SpA is a company belonging to the Italian Finance Office; its mission is to<br />

support e-Government initiatives with particular reference to eProcurement, with the aim to rationalise<br />

public purchasing and improve contracts transparency. Consip has in charge the design and<br />

development of new e-government processes and the related ICT solutions. Its main work consists in<br />

planning the best ideas to implement the strategic directives of Italian Finance Office, acquiring on the<br />

market the best technical solutions and competences to traduce these ideas in ICT applications. This<br />

innovative model permits to obtain two best results: to plan organisational processes for the public sector<br />

well aligned with the Government strategic aims, and to translate projects into practice applying the best<br />

technical solutions on the market.<br />

To implement public eProcurement, Consip uses three main instruments:<br />

Framework agreements;<br />

Electronic auctions and tenders;<br />

eMarketplace.<br />

By framework agreements, Consip awards tenders regarding the purchase of large amount of a well<br />

defined good or service by all the public bodies, with the same prices and conditions drew up in the<br />

contract. Consip is therefore responsible for awarding public contracts or framework agreements for all<br />

the public bodies. In view of the large volumes purchased, this technique helps to increase competition<br />

and streamlines public purchasing. Benefits are important especially for small public agencies, that<br />

purchase small volumes and could access only to local markets with a lower degree of competition.<br />

Moreover, all the public bodies using framework agreements could save the cost to carry out the<br />

agreement by themselves.<br />

By electronic auctions and tenders, a public agency could award a public contract using the technical<br />

platform provided by Consip. This platform complies with the rules drawn up by the Italian laws and<br />

<strong>European</strong> directives about equal treatment and transparency in public purchasing. Therefore the use of<br />

this platform grants the affordability of the technological instruments and prevents public bodies to<br />

develop their own ICT solutions, saving time and money.<br />

The eMarketplace realised by Consip is called MEPA (Mercato Elettronico della Pubblica<br />

Amministrazione/Public Administration Electronic Market). MEPA is a comprehensive and integrate<br />

platform able to manage all the stages of the public procurement: the tenderers selection and ranking, the<br />

48


Clara Benevolo and Renata Paola Dameri<br />

management of tenderers electronic catalogues, the request of quotation or request for order, the final<br />

contract award. On this eMarketplace, public bodies could meet several suppliers and tenderers<br />

previously qualified and therefore compliant with the rules about public purchasing; they can:<br />

Access at a large market and save money thanks to the higher competition between suppliers and<br />

tenderers,<br />

Streamline their procurement activities<br />

Save time thanks to the electronic platform.<br />

By means of the centralisation of public eProcurement by Consip, the Italian Government reached good<br />

goals in a short time. The technological platform for eProcurement is available for all the Italian public<br />

bodies, both small and big ones. This platform permitted to spread the eProcurement practices among all<br />

the public bodies unifying methods and procedures, sparing them the efforts and the expenses related to<br />

the implementation of their own eProcurement application and granting them the affordability and<br />

compliance of the Consip solution. During the period 2004-2010, eProcurement in the Italian public sector<br />

have been increasing from 76 million € to 6000 million €. Among these, about 2286 million € derive from<br />

transactions by Consip (Figure 2).<br />

Figure 2: The Italian public eProcurement volumes (source: Polimi 2010)<br />

Also qualitative benefits have been obtained. The most important benefits regard the process efficiency,<br />

that is, process streamlining, price savings and procurement time reduction. Also the quality of<br />

purchasing has been increasing, especially thanks to the larger number of tenderers and therefore the<br />

higher competition in the market of public procurement (Polimi 2010). Other obtained benefits regard<br />

documents dematerialisation and savings in human resources.<br />

However, the centralised model could not be the best solution for the Italian public eProcurement<br />

practice. Indeed, Italian public bodies and agencies are very heterogeneous and spread throughout the<br />

country. The centralised solution showed several difficulties in involving the smaller bodies and local<br />

administrations. The centralised platform is too generic to respond to the needs of a so large set of<br />

entities, and it is not suitable to meet the expectations and schedule of all the different public bodies, with<br />

their own characteristics and specific needs. Moreover, a centralised solution tends to incorporate best<br />

practices and to maintain them along the time, slowing down the innovation and the research of a better<br />

quality of eProcurement to positively impact on the effectiveness of public services to the citizens. Consip<br />

eProcurement centralised solution was certainly the best answer for the first time adoption, but it is too<br />

much focused on efficiency and it doesn’t take into consideration other aspects.<br />

For all these reasons, a second phase was started, to delocalise eProcurement at regional level and to<br />

drive public agencies to develop their own eProcurement solutions, in order to better reach specific goals<br />

related to their own needs.<br />

4. eProcurement in the Italian public sector: The regional delocalisation<br />

In 2007, the Italian Government issued a law (L. 27 December 2006, n. 296) to rule the adoption of<br />

electronic procurement solutions at the regional level. The Italian State is organised in twenty Regions;<br />

they have administrative autonomy and manage the Regional health system. The aim of this second<br />

49


Clara Benevolo and Renata Paola Dameri<br />

phase of public eProcurement was mainly to overcome the scarce involvement of small and local public<br />

agencies registered in the first phase.<br />

All the Regions, like Consip, have the role of contracting authorities, that is, they can select and authorise<br />

suppliers and award public contracts and framework agreements for all the public bodies and agencies in<br />

their own territory, included the health institutions, compelled to join the agreements.<br />

During the period 2007-2008, the regional contracting authorities carried out 2852 tenders for a total<br />

amount of about 9,7 billions €. 87% of tenders regard goods and services purchasing in the health sector,<br />

mainly in the north and center Italian Regions.<br />

This new phase is today in progress yet. However, it is possible to start to assess this delocalised model<br />

to understand if it is really better than the centralised one. Several aspects should be taken into<br />

consideration:<br />

The aims of this second phase of public eProcurement;<br />

The role of the Regions;<br />

The role played by the State or by another central agency, to govern and coordinate the<br />

delocalisation process.<br />

Unfortunately, the aims and goals of this second phase of eProcurement are not so different respect to<br />

the first phase, and it prevents the qualitative elevation of eProcurement projects. More efficiency of<br />

purchasing, more effectiveness of procurement process, streamlining of administrative activities and<br />

transparency are the main goals, also in this second step. Only two aspects are especially linked with the<br />

delocalisation: the larger participation of both small public bodies and local suppliers. That is, the<br />

enlargement of eProcurement volumes and of the relating benefits. The main goal to reach is ever to<br />

reduce the public spending, and eProcurement is certainly one winning instrument to be applied.<br />

However, the delocalisation has not be used to improve also the quality of public purchasing and the<br />

appropriateness of procurement practices. Also regional eProcurement is too much focused on cost<br />

reduction and administrative streamlining and the importance of the quality of purchasing for the quality of<br />

public services to the citizens is not taken into account.<br />

The role of Regions is the fulcrum of this second phase of public eProcurement. They could play a central<br />

role, especially in:<br />

Spreading the culture of eProcurement on their territory, involving also the smaller public bodies;<br />

Supporting small bodies in implementing eProcurement practices;<br />

Developing specific solutions to support particular needs in their environment;<br />

Studying the impact of eProcurement on the quality of public services delivered to the citizens,<br />

considering also the specific needs of their own area.<br />

At this moment, no surveys have been made, to assess the commitment of Regions and to verify if some<br />

of the roles listed above have been really performed. However, Regions are very different each others<br />

and the implementation of regional eProcurement platforms has been very heterogeneous. In Figure 3<br />

we can see that:<br />

Some Regions implemented eProcurement solutions able to support procurement for all the public<br />

bodies in their own area (the black ones);<br />

Some Regions implemented eProcurement solutions only to support purchasing in the health sector<br />

(the grey ones);<br />

Some Regions are developing during this period eProcurement solutions or have no projects in mind<br />

(the white ones).<br />

It means that each Region started its own eProcurement process, generally without working in<br />

cooperation with other Regions and choosing different ways to implement eProcurement (Avcp 2010).<br />

The further analysis of each regional solution shows that they are very different on behalf on: type of<br />

agency playing the role of contracting agency, type of implemented instruments, degree of participation of<br />

both public bodies and suppliers or tenderers, savings or other benefits obtained by regional<br />

eProcurement. The lack of an evaluation framework to be applied to public eProcurement, based on both<br />

50


Clara Benevolo and Renata Paola Dameri<br />

quantitative and qualitative parameters, prevent to better understand the effectiveness of different<br />

solutions.<br />

realised<br />

work in progress<br />

only for health<br />

Figure 3: The implementation of regional eProcurement<br />

Certainly, the delocalisation phase has partially reached its own goals and it helps to control and to<br />

reduce the public spending, especially in the health sector. More bodies, suppliers and tenderers have<br />

been involved and the competitiveness has increased. Also the administrative processes have been<br />

simplified and sped up.<br />

However, the second phase could be better based on the successes and failures of the first one, to<br />

overcome some problems and especially to aim at more important and large goals, considering also the<br />

quality of public purchasing and not only on the quantity of saved money. The delocalisation is the best<br />

weapon to reach a better alignment between local needs of the citizens, adequate public services and<br />

related purchasing. But to reach these goals two thinks are necessary: the clear definition of objectives to<br />

reach and an evaluation framework to assess if and in which measure the goals have really be reached.<br />

Moreover, in the second phase a central direction is lacked, to coordinate regional initiatives but<br />

especially to share knowledge, competences, best practices and lessons learned. For this reason the<br />

Italian State created a network between Consip and the regional purchasing authorities, aiming to<br />

rationalise the public spending and to create positive synergies in the implementation and use of ICT for<br />

public eProcurement.<br />

Perhaps a knowledge base on public eProcurement could help to better implement effective solutions,<br />

avoiding to repeat errors made by others. The comprehensive analysis of all the regional solutions, of<br />

their characteristics and of their specific goals is useful also to evaluate the trade off between<br />

centralisation and delocalisation. Indeed, to organise eProcurement at the regional level on one side<br />

improves the appropriateness of solutions and the involvement of local bodies; on the other side, it<br />

reduces the volumes of purchasing and therefore the economy of scale deriving from the centralised<br />

procurement. Only having well clear the desired results and behaviours it is possible to choice between<br />

the higher efficiency of the centralised model and the better specificity and adequacy of the delocalised<br />

one.<br />

5. Conclusions<br />

In Italy, public bodies are about 11.000. Among them, about 50% use eProcurement, but only a few<br />

hundreds are mature and intensive users (Polimi 2010). The more used instruments are eMarketplaces<br />

and online stores, auctions and framework agreements are less used. The savings are estimated<br />

between 10 and 50% of the electronic spending.<br />

Therefore, eProcurement in Italy could be considered a successful initiative, especially regarding the<br />

amount and volume of public purchasing, but it is not enough spread. The main role is played by Consip,<br />

that carries out 50% of purchasing; the prevalent model is hence the centralised one. The delocalised<br />

model have been starting, but with very heterogeneous results. Several regional purchasing authorities<br />

51


Clara Benevolo and Renata Paola Dameri<br />

are developing and implementing eProcurement solutions, but only nine regional eProcurement platforms<br />

are really working and only four Regions appear mature in the use of eProcurement. Moreover,<br />

procurement processes have not been redesigned and benefits are gained only on efficiency and<br />

spending; no evidences are available, regarding the actual impact of eProcurement on the quality of<br />

public services for the citizens.<br />

Examining the Regional implementation of eProcurement, some evidences are useful, to plan<br />

improvements to be spread all over the State.<br />

Only 4% of purchasing is carried out on the electronic channel; this number could be increased,<br />

especially regarding electronic auctions and framework agreements.<br />

The commitment of local authorities is crucial for the success of eProcurement initiatives, such as the<br />

role of central and regional purchasing authorities in supporting and promoting the use of<br />

eProcurement by smaller public bodies.<br />

To improve the efficiency of e-purchasing, a more effective requirements planning is necessary; it<br />

could permit to optimise quantity and quality of purchasing. No evidences are available, of<br />

requirements planning practices and instruments to support eProcurement.<br />

No evaluation frameworks are available, to quantify savings in money and time and to award the best<br />

public bodies and managers in eProcurement implementation. Moreover, it is necessary to involve<br />

the public opinion in the assessment of eProcurement and its actual or potential impact on public<br />

services. EProcurement results should be measured, evaluated respect to politic and administrative<br />

goals of each public body and disclosed to citizens and communities using the most adequate media<br />

and instruments.<br />

The topic of public eProcurement is therefore actual and it promises future interesting returns, if<br />

adequately studied in depth. Further researches could be useful, especially regarding the following<br />

aspects.<br />

To analyse some best cases, regarding public eProcurement in the Italian Regions. We discovered<br />

two or three successful platforms supported by best practices in process redesign, requirements<br />

planning and involvement of smaller public bodies. The case studies could be the basis for defining a<br />

general framework for public eProcurement implementation in local purchasing authorities.<br />

To better analyse the actual impact of public eProcurement on the efficiency of purchasing activities<br />

and practices, and on savings deriving from lower prices and higher competition among suppliers and<br />

tenderers. Till now, such surveys have been conducted only using panels and interviews and<br />

qualitative indicators. No affordable quantitative metrics are available. Moreover, the survey could<br />

better detailed the research, distinguishing between different techniques used for eProcurement<br />

(auctions, tenders, marketplaces, e-stores) to find relations between techniques and benefits, to<br />

answer to different purchasing requirements.<br />

To design a comprehensive evaluation method, able to assess not only the savings in time, money<br />

and human resources deriving from the process streamlining, but also to evaluate the impact of best<br />

eProcurement practices on the quality of public services for citizens; such an evaluation framework<br />

should include also a disclosure framework to involve communities in the assessment of<br />

eProcurement initiatives and their impact on the amount and quality of public spending.<br />

References<br />

Alsac, U. (2007), “Use of eProcurement in Turkey’s public health sector”, Journal of Public Procurement, Vol. 7, No.<br />

3, pp. 333-361.<br />

Ash, C. and Burn, J. (2003) “Assessing the benefits from e-business transformation through effective enterprise<br />

management”, Journal of Information Systems, 12, pp. 297-308.<br />

Avcp (2010) (Autorità per la Vigilanza sui Contratti Pubblici di Lavori, Servizi e Forniture) Censimento ed analisi<br />

dell’attività contrattuale svolta nel biennio 2007-08 dalle Centrali di Committenza Regionali e verifica dello stato<br />

di attuazione del sistema a rete, [online], www.avcp.it.<br />

Bruno, G., Esposito, E., Mastroianni, M. and Vellutino, D. (2005) “Analysis of public eProcurement web site<br />

accessibility”, Journal of Public Procurement, Vol. 5, No. 3, pp. 344-366.<br />

Cabras, I. (2010) “Use of EProcurement in Local Authorities' Purchasing and Its Effects on Local Economies:<br />

Evidence from Cumbria, UK”, <strong>European</strong> Planning Studies, Vol. 18, No. 7, p. 1133.<br />

Carpineti, L., Piga, G. and Zanza, M. (2006), The Variety of Procurement Practice: Evidence from Public<br />

Procurement, in Dimitri, N., Piga, G. and Spagnolo, G., eds., Handbook of Procurement, Cambridge University<br />

Press.<br />

52


Clara Benevolo and Renata Paola Dameri<br />

Croom, S. and Brandon-Jones, A. (2005) “Key issues in eProcurement: procurement implementation and operation<br />

in the public sector”, Journal of Public Procurement, Vol. 5, No. 3, pp. 367-387.<br />

Dooley, K. and Purchase, S. (2006) “Factors influencing eProcurement usage”, Journal of Public Procurement, Vol.<br />

6, No. 1/2, pp. 28-45.<br />

Hommen, L. and Rolfstam, M. (2009) “Public procurement and innovation: towards a taxonomy”, Journal of Public<br />

Procurement, Vol. 9, No. 1, pp. 17-56.<br />

MacManus, S. (2002) “Understanding the incremental nature of eProcurement implementation at the state and local<br />

level”, Journal of Public Procurement, Vol. 2, No. 1, pp. 5-28.<br />

Mishra, A., Konana, P. and Barua A. (2007) “Antecedents and Consequences of Internet Use in Procurement: An<br />

Empirical Investigation of U.S. Manufacturing Firms”, Information Systems Research, Vol. 18, No. 1, pp. 103-<br />

122.<br />

Moon, J. (2005) “EProcurement management in state governments: diffusion of eProcurement practices and its<br />

determinants”, Journal of Public Procurement, Vol. 5, No. 1, pp. 54-72.<br />

Polimi (2009) Innovazione ed efficienza nella Pubblica Amministrazione: il ruolo dell’eProcurement, Osservatorio<br />

eProcurement nella Pubblica Amministrazione, Politecnico di Milano – Dipartimento di Ingegneria Gestionale.<br />

[online], www.osservatori.net.<br />

Polimi (2010) eProcurement: una concreta via verso innovazione, efficienza e trasparenza per la PA italiana,<br />

Osservatorio eProcurement nella Pubblica Amministrazione, Politecnico di Milano – Dipartimento di Ingegneria<br />

Gestionale. [online], www.osservatori.net.<br />

Reddick, C. (2004) “The growth of eProcurement in American state governments: a model and empirical evidence”,<br />

Journal of Public Procurement, Vol. 4, No. 2, pp. 151-176.<br />

Rendon, R. (2008) “Procurement process maturity: key to performance measurement”, Journal of Public<br />

Procurement, Vol. 8, No. 2, pp. 200-214.<br />

Singer, M., Konstantinidis, G., Roubik, E. and Beffermann, E. (2009) ”Does eProcurement save the state money?“,<br />

Journal of Public Procurement, Vol. 9, No. 1, pp. 58-78.<br />

Somasundaram, R. (2004) "Diffusion of eProcurement in the Public Sector: Revisiting Centralized vs. Decentralized<br />

Debates As a Twist in the Tale", Proceedings of the 13 th <strong>European</strong> <strong>Conference</strong> on Information Systems, Turku,<br />

Finland, June.<br />

Tavi, J. (2008) “Learning From Global World-Class eProcurement Practices”, Strategic Finance, Apr., Vol. 89, No.<br />

10, pp. 24-30.<br />

Vaidya, K, Sajeev, A. and Callender, G. (2006) “Critical factors that influence eProcurement implementation success<br />

in the public sector”, Journal of Public Procurement, Vol. 6, No. 1/2, pp. 70-99.<br />

Walker, H. and Harland, C. (2008) “EProcurement in the United Nations: influences, issues and impact”, International<br />

Journal of Operations & Production Management, Vol. 28, No. 9, pp. 831-857.<br />

Wirtz, B., Sebastian, L. and Schierz, P. (2010) ”An Empirical Analysis of the Acceptance of EProcurement in the<br />

German Public Sector“, International Journal of Public Administration, Vol. 33, No. 1, pp. 26-42.<br />

53


ICTs as Weapons of Mass Interaction - Motivations and<br />

Implications of Mediated Control<br />

Kofi Agyenim Boateng<br />

British Institue of Technology & E-commerce, London, UK<br />

boateng@bite.ac.uk<br />

fiboat@yahoo.com<br />

Abstract: Interaction has long been an integral aspect of organisation’s life. Hence, in recent times interactions<br />

driven by Information Communications Technologies (ICT) have gained significant acceptance and momentum in<br />

contemporary organisational settings. This is demonstrated by their enormous adoption and wide-ranging<br />

employment across the various levels of an organisation’s hierarchy. Consequently, businesses have started to<br />

expand their investment in, and broaden the frontiers of, technology-mediated interaction. This has meant the<br />

involvement of such communication devices as mobile phones, personal digital assistants (PDA), BlackBerries,<br />

laptops, as well as other forms of portable and immovable computing technologies to provide enduring support for<br />

both enterprise-wide and external interactions. This study adds to the existing conceptualisations of technology<br />

mediated communication by delving into the fundamental rationalities, manifestations and implications of mediated<br />

control. To this end, an exploratory study is conducted in a moderately large UK food company that distributes semiprepared<br />

food to small restaurants, and examines the applications of both mobile and stationary technology tools for<br />

undertaking different enterprise-wide communication endeavours under dynamic work strategising commitments.<br />

The study reveals that first, the application of multiplicity of ICT-driven interaction tools generate diverse social<br />

implications, both pleasant and unpleasant to the business organisation. Second, multiple actors have different<br />

motivations for the application of varied technology tools in the accomplishment of a given organisational tasks,<br />

however similar such tasks may be. Finally, study as a result provides significant managerial and theoretical insights<br />

and implications for organisational interaction and the distribution and service delivery businesses in the briskly<br />

growing digital market.<br />

Keywords: ICT-driven interaction, social presence theory, media richness theory, emotion<br />

1. Prolegomenon<br />

The application of Information Communication Technologies (ICTs) in the many facets of organisational<br />

interaction is quite a recent phenomenon in spite of the fact interaction has always been an instrumental<br />

element of the life organisations. Business enterprises continue to expand their outlay in ICTs to reap the<br />

benefits afforded by technology mediated interaction in its varied breeds, such as web-driven media,<br />

mobile, fixed telephones, and short message service (SMS), among others (Bernett et al., 2002). An<br />

appealing aspect of technology mediated interaction is the flexibility and the vitality particularly associated<br />

with the many modes of its expression (Olson and Olson, 2000 p.141). A significant aspect of technology<br />

mediated interaction is the dynamism chiefly connected with the many avenues of its expression. Part of<br />

their flexible nature is demonstrated by their capacity of establishing interaction and exchange relations<br />

with others not collocated; what can conveniently be described as ‘virtual collaboration’(Olson and Olson,<br />

2000 p.141).<br />

Previous studies on technology mediated interactions have attended to the material features of the<br />

applied media (Daft and Lengel, 1986a), social considerations (Yoo and Alavi, 2001) reconstruction of<br />

interaction (Ngwenyama and Lee, 1997) and aspects of textual configurations (Orlikowski and Schultze,<br />

2004). While admitting that these studies made monumental impact on the mainstream discussion on<br />

ICT-driven interaction, they often overlook the emotional repercussions and the underlying psychological<br />

rationalities that come to the fore in such interactions. Still, the studies that intimately address the social<br />

elements of ICT interaction (Suchman, 1987) to tease out the emotional consequences that the<br />

interlocutors of ICT interaction experience. This paper wishes to broaden these investigations by having<br />

recourse to the theories of social presence and media richness and thereby advance the idea of how<br />

making sense of these ICT tools can diverge even across the people doing similar job within the same<br />

organisation.<br />

In particular, the paper advances the idea that ICT interaction can equally generate ‘e-ritual’ chains that<br />

are fundamental in getting agents undertake most social activities just as face to face interaction. We<br />

explore this concept by bringing into close scrutiny the limited perspective of Collins’ (2004) interaction<br />

ritual chains. This objective is achieved by embarking on a longitudinal study in a medium-size private<br />

company that specializes in supplying semi-prepared food to fast-food shops in and around London. The<br />

study seeks to address the following research queries:<br />

54


Kofi Agyenim Boateng<br />

To what extent can we better understand the rationalities and motivations that drive technology<br />

mediated interaction in collaborative environment?<br />

How does the use of ICT interaction tools affect the emotional make up of people?<br />

This paper resumes with a cursory look on the mainstream views on interaction with a touch on both<br />

media richness theory and social presence theory. The methodological approach to the study follows and<br />

the results from the field of investigation are immediately provided in that order. Eventually, the paper<br />

analyses the implications for both theoretical and practical standpoints and conclude by suggesting<br />

prospects for future areas of research.<br />

1.1 Interaction –contemporary considerations<br />

Interaction has witnessed many path-breaking studies and the masterpiece by Gofmann (1967) provides<br />

the foundation for many branches on the subject matter. Interaction has been conceptualised<br />

fundamentally in its mutual sense and therefore seeks to establish ‘reciprocal events that require at least<br />

two objects and two actions. Interaction occurs when these objects and events mutually influence one<br />

another’(Wagner, 1994 p. 8). The mutual aspect of interaction can take many possible forms such as<br />

collaboration, co-operation and coordination, all within the realm of communication. Interaction then,<br />

enables the establishment of networks for reaching out to contacts without necessarily their body<br />

presence (Peters, 1999) by means of the mediating capabilities of ICTs.<br />

ICT-driven interaction generates social relationships that form the basis of community integration<br />

(Calhoun, 1992) as various degrees of social relationships are likely to emerge from the ongoing<br />

exchanges. As such, human behaviour has impacted and informed the design, architecture and<br />

functionalities of these technologies. This mutual shaping between technology and human agency is<br />

observed elsewhere with the claim that ‘whatever effect technology may have on social forms is heavily<br />

shaped by the social, cultural and institutional relations in which the technology is embedded’ (Kallinikos,<br />

2007 pp. 278 - 279).<br />

The application and use of mediation tools in getting in touch with one’s interaction partner broadens the<br />

meaning and scope of interaction as the mediating capabilities of the technology. A capability that allows<br />

for ‘mutual tuning-in relationship’ (Schutz, 1971) based on mediated interaction, because it is predicated<br />

on shared dependency, common identity, mutual experience and joint context that is captured in a<br />

collective purpose directed toward common objectives (Schmidt and Bannon, 1992). This view suggests<br />

that interaction leads people to initiate strategic moves, develop routines, cultivate practices and<br />

procedures that give birth to organisations and provide the means for their vital sustenance (Hughes,<br />

1971). Within the spirit of symbolic interactionist (Blumer, 1969), the enduring nature of interaction does<br />

not only subsist in a context but it also constructs the context.<br />

It therefore comes as with a little surprise that Kakihara and Sørensen (2002) mainly link the building<br />

process of the phenomenon of knowledge (including an appreciation of the social backgrounds of the<br />

interactants) with the idea of interaction. The reason for this, the pair claim, is that the emergence of<br />

knowledge traces its sources from interaction. And this interaction forms the very rationale of complex<br />

exchanges inherently meshed in human society.<br />

Mediated or ICT-driven interaction recognises the use of technical elements to function as the platform on<br />

which interaction can be consummated. Thompson (1995 p. 83) recollects that technical medium like<br />

“paper, electrical wires, electromagnetic waves, etc.” facilitates the transmission of symbolic and<br />

information contents to parties who are either disconnected by time or space or occasionally both. The<br />

application of technical objects capable of setting two parties to engage in meaningful interaction<br />

underlies the concept of both media richness and social presence theories, the two of which produce<br />

different interaction effects for the interacting people. Nonetheless, mediated interaction affords the<br />

interactants very limited symbolic tokens for the reduction of uncertainty.<br />

1.2 Media richness and social presence theories – a case of mediated interaction<br />

Social presence holds the notion that mediated interaction has the potential to stimulate the features and<br />

benefits of face-to-face interaction Short, Williams et al (1976) and emphasises the extent to which one<br />

senses the immediate presence and significance of his or her interaction partner (Kayany et al., 1996).<br />

Social presence is crucial for building our knowledge and perception of other people’s behaviour and<br />

attitudes. Social presence is usually judged by the differences in meaning that borders on emotions or the<br />

55


Kofi Agyenim Boateng<br />

suitability of different media for a variety of similar organisational communication tasks (Argyle, 1969,<br />

Birdwhistle, 1970).<br />

The source of the media richness is directly connected to Daft and associates (1986b, 1987) when they<br />

made investigations into managers’ inclination to apply oral type of communication even when its<br />

mediated form is reasonable. The media richness theory suggests the ability of communicated<br />

information via a medium to limit the extent of equivocality on four counts, namely, the ‘speed of<br />

feedback, cue multiplicity, language variety, and personal focus’ (George and Carlson, 1999 p. 3). Key to<br />

media richness is the hypothesis that there is a disparity between jobs and specific characteristics;<br />

various media differ in the degree to which they can dispatch the kinds of information needed and the<br />

heterogeneous task characteristics.<br />

Consistent with the research questions outlined above, we conducted a field study to investigate how<br />

emotions and different psychological rationalities become evident in individuals making use of ICT<br />

interaction tools.<br />

2. Method<br />

The section addresses the ways and means of the whole research effort. The procedural method is<br />

fundamental for outlining the scheme for grounding the various parts and procedures of the entire study<br />

(Grunow, 1995). In short, the methodology goes to a greater length at laying down a sturdy and robust<br />

validity of the outcome of the research undertaking (Yin, 2003).<br />

The path to this methodological examination is followed with the aid of the three primary mechanics of<br />

human perception, namely, rationalism, empiricism and faith (Thieme, 2003). Rationalism identifies reality<br />

by means of reason. Empiricism does so through a support on the senses of sight, touch, taste, smell<br />

and sound. And faith defines reality via the confidence in the authority of another entity. Using all three<br />

systems of perception in the data-collection process would be beneficial in the sense that ‘rational proof<br />

and instruction do not fully exhaust the sphere of knowledge’ (Gadamer, 2004 p. 21). Therefore<br />

appropriating all three systems in a methodological process would ensure a demonstration of consistent<br />

and logical thinking thereby effectively limiting the researcher from going arbitrary with his judgements; so<br />

that the conclusions would not be treated ‘as a figment of my imagination’(Thieme, 2003 p. 2).<br />

To this end, the following data collection techniques were adopted for this study. The aim is to better<br />

appreciate phenomena-in-context inquiries against the background of certain professionals’ in their<br />

natural, operationally distinct but electronically connected environment and their cross-subjective<br />

predilections, this work mainly used qualitative data.<br />

2.1 Direct observation and interview<br />

The use of direct observation method allows the researcher to focus on the aspects of the research that<br />

are of particular interest to him without the possibility of being biased or influenced by ‘second-hand’<br />

information from the participants of the study. This necessarily occasioned long hours of sitting beside<br />

telesales reps, customer service reps, delivery drivers and team leaders as they handle both incoming<br />

and outgoing calls. There other situations where I was privileged enough to have been given the<br />

opportunity by a telesales rep to listen in to incoming telephone conversations from customers. I should<br />

admit that this ‘privilege’ was not officially permitted by top management; it was nonetheless made<br />

possible as a result of tactful arrangement I struck with a particular telesales rep. I also had the chance to<br />

travel to most delivery areas in the South-East of London with some of the delivery drivers.<br />

Most of these interviews lasted about 50minutes and some crossed the hour mark due to telephone<br />

interruptions from customers and work colleagues. Details from these interviews were electronically<br />

recorded and some notes also taken to fill data gaps during the transcription. Notes were read to the<br />

interviewees after every interview to allow confirmation and corrections to guarantee validity.<br />

2.2 Miscellaneous evidence<br />

Other forms of documentation such as quarterly progress meeting reports, company brochures,<br />

promotion leaflets, and many others were used to understand and make sense of the data that emerged<br />

from the interviews. The pieces of information gleaned from these materials proved quite helpful in later<br />

analysis and discussion of the data.<br />

56


Kofi Agyenim Boateng<br />

3. Findings –interaction activities at Adom Food Services Ltd<br />

Adom Food Services Ltd., henceforth designated as AFSL, trades in semi-prepared food and other food<br />

related products to small and medium-size restaurants in and around London. It is appropriate to point<br />

out at this stage that Adom is a fictitious nomenclature intended to protect the real identity of organisation<br />

where this study was carried out. AFSL has lately, through the use of technology, automated greater<br />

parts of the processes involved in commercial exchanges with its customers from order processing to<br />

delivery. Two unique services are provided: order processing by telesales representatives and order<br />

delivery by delivery drivers. The telesales reps are made up of about 12 different teams of roughly 5<br />

members, headed by a team leader each, while delivery is undertaken by a fleet of approximately 200<br />

vehicles.<br />

The study is contextualised around order processing and delivery activities which conspicuously feature<br />

telesales reps and delivery drivers in their use of portable, mobile and stationary ICTs in performing their<br />

assigned duties. The decision to use telesales reps and delivery drivers as the unit of analysis of this<br />

thesis is due to their particular use, and heavy reliance on ICTs to drive interaction in all its stripes in the<br />

enterprise-wide endeavours at AFSL. Secondly, the resolve to concentrate on ICT-driven interaction<br />

takes into due consideration the issues emotional repercussions that are significant in demarcating the<br />

context of this study.<br />

AFSL serves an extensive mix of customers including wholesalers and retailers. These include cafés,<br />

restaurants, burger bars, football clubs, ‘chicken and chips’ shops among others. The commercial<br />

activities at AFSL go almost seamlessly on a 24-hour basis from Monday to Friday and, up to noon on<br />

Saturdays and, sometimes Sundays. Supplying semi-packaged foods and drinks to small restaurants and<br />

shops requires registered customers and potential customers to telephone an all-purpose telephone<br />

number anytime between 9am and 5pm from Monday to Friday and, occasionally on Saturdays to make<br />

their orders. The details of the orders are then stored on a central database of the company until after<br />

5:30 p.m., when they have been matched against their respective contact details of the customers who<br />

ordered them.<br />

Building shipment resumes after all route planning activities have been done. Night loading is performed<br />

soon afterwards. Loading is done all night until about 6 a.m., by which time the individual drivers of the<br />

various trucks have arrived at the company’s premises.<br />

Having prefaced the diverse operational manifestations at AFSL, it is appropriate to bring out some of the<br />

details of the telephone conversations that emerged from the study. The thoughts from the interviewees<br />

(telesales reps, delivery drivers, telesales team leaders) reflect different philosophical orientations on the<br />

use of ICTs in mediated interactions. These orientations inform two broad thematic pathways, namely,<br />

emotional and psychological consequences on one hand and the motivational rationalities underlying the<br />

application of ICT-driven interactions. It must be said that all the names mentioned do not reflect the true<br />

identity of the people interviewed.<br />

One telesales rep expressed her frustration in promoting and selling on the phone to customers thus:<br />

‘When you’ve got like 20 calls holding and you know you gonna take the next call the<br />

customer is gonna be angry because they’ve been waiting, you know, and they don’t wanna<br />

hear you trying to sell them something; it’s hard’ – Theresa<br />

Another made the following remark.<br />

‘They [team leaders] are always on your back, I can’t force 11 people to get new lines’ –<br />

Georgina.<br />

‘Personally I don’t think I really like my job [a reference to the telesales work] anymore, I<br />

have got commitments so I have to come and get paid’ – Florence.<br />

‘I spoke to a customer who asked me to marry him over the phone … he went like ‘alright<br />

darling’, ‘okay sweetheart’ all this sort of talk’ – Louisa.<br />

‘You can sell a new line (in other words a new product) easier to a man than to sell to a<br />

woman… women, you know, just get on a bit but men oh!...they hear a woman’s voice and<br />

they know she is young and they go like: What’s your name, darling? Where’re you from? ...<br />

and stuff like this’ – Frances.<br />

57


Kofi Agyenim Boateng<br />

From the viewpoint of the delivery drivers, the following thoughts came to the fore during the interviews.<br />

A delivery driver’s reason for maintaining close relationship with a customer of the company is captured<br />

in the following way.<br />

‘My friend, the office people don’t know that this customer has my mobile number. I talk to<br />

her everyday and she knows when I am going to make deliveries. Sometimes she calls to<br />

find out when I am going to drop her order. And I say 10 minutes, half hour or even one<br />

hour, you understand? Me and this customer we are very good friends for a long time now’ –<br />

Gordon.<br />

In connection with the rationalities and motivational factors underpinning the application and preference<br />

of one technology tool to another the following matters surfaced during the interviews.<br />

‘Personally the email is a second option. I prefer the use of the telephone because that gives<br />

me a more appropriate description to the problems the staff may have. I don’t get immediate<br />

response to emails and I can’t put them any questions for instant response. That means I<br />

can’t really fish out what the problem is when I use the email’ – Thomas.<br />

‘Using emails wastes a lot of time, as talking on the phone gives me the opportunity to get<br />

instant feedback for my queries’ – Rosina.<br />

‘Talking over the telephone, you never know if the other person is listening to you so when<br />

you have the eye contact it makes a huge difference’ – Daniels.<br />

‘I send an email to the field sales reps and all of them get the same message<br />

simultaneously, it is not misinterpreted, and it is accurate, but if it is a telephone (mobile or<br />

landline), I have to send it one after the other. However, if you speak to someone, two or<br />

three different people asking the same question, you’d get two or three different<br />

interpretations…so that [an allusion to email] works very well for me’ – Robinson.<br />

One team leader expressed an innovative application of the email in enriching mediated interaction.<br />

According to him, he could use colour codes to highlight certain aspects of communicated messages.<br />

The research had the opportunity to witness an instance of colour-enriched mails.<br />

‘The beauty of email is you can prioritise your work, set reminders, colour code in terms of<br />

priority, arrange and delegate to other people’ – Katlin.<br />

The next section presents an analysis of the paper to flesh out the two principle themes of emotional<br />

consequences of mediated interaction and motivational factors triggering certain aspects of technology<br />

mediated interaction.<br />

3.1 ICT interaction and emotional dynamics<br />

ICT-driven interaction can result in certain psychological and emotional challenges that may impact<br />

negatively or positively on the performance of the individual concerned, especially if such emotional<br />

surges are not well handled by the people involved. It is obvious from the study that interactions<br />

empowered by ICTs have assumed nearly the same mechanical dimension as face-to-face interactions.<br />

The affordance by these technology objects makes it possible for what Collins (2004) refers to as<br />

‘emotional energy’ to be charged, the effect of which can match up to many close, proximal<br />

communication. Thus ICTs permit the essential logistical support for the breeding of networks, described<br />

as ‘networked individualism’ (Castells, 2001 p 130) as a central mechanism for flourished communication<br />

relationships. It can be stressed – without equivocation – the point that ICTs appear to have assumed a<br />

dominant role in configuring complex network of social organisation. These networks offer a gateway for<br />

individuals to share communication principles both ‘on-line and off-line, on the basis of their interests,<br />

values, affinities, and projects’ (Castells, 2001 p. 131).<br />

Proponents of the doctrine of social presence claim that face-to-face interaction can hardly be argued to<br />

have a perfect substitute (Culnan and Markus, 1987, Walther, 1992). However, it is evident from the<br />

study how repeated mediated interactions breed familiarity. This is found in the remarkable revelation of<br />

how ICT-driven interaction inspires, strengthens and sometimes affects the configuration of social<br />

relationships. The interview discovered breaches of strict organisational codes of practice as a result of<br />

familiarity between telesales reps, delivery drivers and customers occasioned by repeated and intensive<br />

mediated patterns of interaction.<br />

58


Kofi Agyenim Boateng<br />

The study for instance learned informally, if confidential, how a telesales rep developed amorous<br />

relationship with a customer of the organisation. He was promptly sacked when this ‘unofficial rapport’<br />

became known to the organisation. Therefore the material properties afforded by the technology enables<br />

what can take place face-to-face to equally prevail in a communication at a distance. This could be as a<br />

result of the observation elsewhere that in mediated interaction<br />

“there have to be signs, objects of some sort that are about some thing, objects whose<br />

function is reference rather than presence” (Borgmann, 1999 p.17).<br />

Hence mediated interaction makes it possible for initial encounters at the point of service delivery to<br />

develop into relationships paved by a bond of trust that follows dynamic interactions between the social<br />

agents. Some of the experiences of the telesales reps adequately capture this viewpoint. The discussion<br />

on this segment cannot be complete without recollecting certain responses from two telesales reps about<br />

product promotion on the phone which was laden with worry, frustration and intimidation. Indeed, one<br />

telesales rep could hardly suppressed her feelings when she declared in an interview that she feels quite<br />

apprehensive going to work at the thought of the challenge in meeting the daily sales target via the<br />

telephone. This worryingly frustrating experience connects well with an observation elsewhere that ‘the<br />

increasing competiveness of the modern economy has rendered people slaves to their workplaces’<br />

(Bolchover, 2005 p. 2). Also, Bunting’s (2005) ‘Willing Slaves…’ similarly has a remarkable resemblance<br />

with a workplace characterised by a stressed and overworked staff.<br />

3.2 Manifest rationalities in ICT-driven interaction<br />

Varying degrees of answers are understood to have motivated the application of various forms of<br />

communication tools for interaction predominantly, among the telesales reps. These declarations imply<br />

that user appropriation of ICTs depends on different situational factors connected with multiple rational<br />

explanations (Franz and Robey, 1994).The attribution of personal factors by the category of people<br />

identified in the study reflects sense-making posture in technology use that dwells largely on personal<br />

experience rather than any other social consideration. Different telesales reps assign different reasons for<br />

the use of the same technology device for almost the same job.<br />

It can also be conveniently argued that with respect to the manifest rationalities in ICT interaction, the<br />

technology opens up opportunities for manoeuvrability aimed at the achievement of other personal goals.<br />

For instance, colour coding email message to make it meaningfully rich and real, setting priorities of<br />

communicative trends with colleagues and the ability to transmit across spatial context makes email an<br />

appealing choice for engaging in certain aspects of mediated interaction.<br />

4. Concluding remarks<br />

Specific portions of the findings of this paper are unsympathetic to, and occasionally undermine, the<br />

fundamental postulates of the social presence theory. For instance, some interviewees showed they pay<br />

least regard to matters of social presence when instantiating interaction with their working colleagues.<br />

They have a firm indifference to the use of ICTs with specific reference to telephone and email for<br />

mediating their interaction. Their position seems to depend not essentially on the device being adopted<br />

but on the content of the message being broadcast via the device. This reflection was echoed by a<br />

couple of telesales reps’ opinions for which the following statement is illustrative.<br />

‘Just the same thing that I would say to the other person on the line is what I put down on the<br />

email; if I want to say something I would say the same on the email as I would via the phone’<br />

Collins (2004) largely sees emotions as a two-edged piece for erecting and changing social structures on<br />

the one hand, and a potential source of societal disagreements on the other, through the use of what he<br />

refers to as ‘emotional energy’. Mediated ICTs can build emotional energy through what I call ‘electronic<br />

ritual chains’ from the concept of Collins understanding of the ‘ritual’ in face-to-face encounters. This<br />

electronic means of building emotions by mediated technology needs to be taken further in all forms of<br />

communication at a distance. The emotional conditions necessitated by those communication activities<br />

are worthy of reflective consideration. For instance, evidence of emotional spillover in ICT-driven<br />

interaction arose from two divergent sources; one within the company (internal) and the other from<br />

outside the organisation (external).<br />

As emotions assumes the better part of a person, they are likely to become subjective to the degree that<br />

they tend, even if briefly, overlook their immediate allegiance to the cause of the organisation as a group<br />

committed to a common cause. A near quarreling remark by a telesales rep summarises the collective<br />

59


Kofi Agyenim Boateng<br />

effect of employee disenchantment when emotions surge to an intolerable level as a result of mediated<br />

interaction. This is because when pleasure, excitement and even enjoyment gain prominence in everyday<br />

interactions rising from intense communication, the possibility of getting involved personally, and with<br />

one’s emotions, can be unexpectedly sharp.<br />

Therefore our study of ICT-driven interactions should not limit our horizon to the artefacts and possibly<br />

the social implications of their implication into everyday life. Studies are encouraged to look intimately at<br />

how the behaviour of people changes overtime and the consequences and dynamics of such<br />

transformation. Again, this study opens the floodgate for future research to go beyond logistics to say<br />

emergency services, e.g. ambulance, the police or the fire service as a means for informing both theory<br />

and practice in ICT-driven interactions.<br />

References<br />

Argyle, M. (1969) Social Interaction, Methuen, London.<br />

Bernett, H. G., Masi, D. M. and Fischer, M. J. (2002), Web-enabled call centres: a progress report. Business<br />

Communications Review, 32, 7, 38-43<br />

Birdwhistle, R. L. (1970) Kinesics and Context, University of Philadelphia Press, Philadelphia.<br />

Blumer, H. (1969) Symbolic Interactionism: Perspective and method, Prentice Hall, Englewood Cliffs, NJ.<br />

Bolchover, D. (2005) The Living Dead: Switched Off, Zoned Out - The Shocking Truth About Office Lice, Capstone,<br />

Chichester, West Sussex.<br />

Borgmann, A. (1999) Holding On to Reality: The Nature of Information at the Turn of the Millennium, The University<br />

of Chicago Press, Chicago.<br />

Bunting, M. (2005) Willing Slaves: How the Overwork Culture is Ruling Our Lives, HarperPerennial, London.<br />

Calhoun, C. (1992), The infrastructure of modernity: Indirect social relationships, information technology and social<br />

integration In Social change and modernity (Eds, Haferkamp, H. and Smelser, N.) University of California Press,<br />

Berkeley, pp. 205-236.<br />

Castells, M. (2001) The Internet Galaxy: Reflections on the Internet, Business, and Society, Oxford University Press,<br />

Oxford.<br />

Collins, R. (2004) Interaction Ritual Chains, Princeton University Press, Princeton, NJ.<br />

Culnan, M. J. and Markus, M. L. (1987), Information Technologies In Handbook of Organizational Communication:<br />

An Interdisciplinary Perspective<br />

(Eds, Jablin, F. M., Putnam, L. L., Roberts, K. and Porter, L.) Sage, Newbury Park.<br />

Daft, R. L. and Lengel, R. H. (1986a), Organisational information requirements, media richness and structural design.<br />

Management Science, 32, 5, 554-571<br />

--- (1986b), Organizational Information Requirements, Media Richness, and Structural Design. Management<br />

Science, 32, 5, 554-571<br />

Daft, R. L., Lengel, R. H. and Trevino, L. K. (1987), Message Equivocality, Media Selection, and Manager<br />

Performance: Implications for Information Systems. MIS Quarterly, 11, 3, 355-366<br />

Franz, C. R. and Robey, D. (1994), An Investigation of User-Led System Design: Rational and Political Perspectives.<br />

Communications of the ACM, 27, 12, 1202-1209<br />

Gadamer, H.-G. (2004) Truth and Method, Continuum, London.<br />

George, J. F. and Carlson, J. R. (1999) In Proceedings of the 32nd Hawaii International <strong>Conference</strong> on System<br />

Sciences, Vol. 1 IEEE Computer Society Washington, DC, USA, Hawaii<br />

Grunow, D. (1995), The Research Design in Organization Studies: Problems and Prospects. Organization Science,<br />

6, 1, 93-103<br />

Hughes, E. C. (1971) The Sociological Eye: Selected Papers On Institution and Race (Part I) and Self and the Study<br />

of Society (Part II), Aldine Atherton, Chicago.<br />

Kakihara, M. and Sorensen, C. (2002), Exploring Knowledge Emergence: From Chaos to Organizational Knowledge<br />

Journal for Global Information Technology Management, 5, 1-14<br />

Kallinikos, Y. (2007), ICT, organizations, and networks In The Oxford Handbook of Information and Communication<br />

Technologies (Eds, Mansell, R., Avgerou, C., Quah, D. and Silverstone, R.) Oxford University Press, Oxford,<br />

pp. 273-292.<br />

Kayany, J. M., Wotring, E. C. and Forrest, E. J. (1996), Relational Control and Interactive Media Choice in<br />

Technology-Mediated Communication Situations. Human Communication Research, 22, 3, 399-421<br />

Ngwenyama, O. K. and Lee, A. (1997), Communication Richness in Electronic Mail: Critical Social Theory and the<br />

Contextuality of Meaning. MIS Quarterly, 21, 2, 145-167<br />

Olson, G. M. and Olson, J. S. (2000), Distance Matters. Human-Computer Interaction, 15, 139-178<br />

Orlikowski, W. J. and Schultze, U. (2004), A practice perspective on technology-mediated network relations: the use<br />

of Internet-based self-serve technologies. Information Systems Research, 15, 1, 87-106<br />

Peters, J. D. (1999) Speaking into the air: A history of the idea of communication, University of Chicago Press,<br />

Chicago.<br />

Schmidt, K. and Bannon, L. (1992), Taking CSCW Seriously: Supporting Articulation Work. Computer Supported<br />

Cooperative Work (CSCW), 1, 7-40<br />

Schutz, A. (1971), Making Music Together In Alfred Schutz, Collected Papers II: Studies in Social Theory (Ed,<br />

Brodes, A.) Martinus Nijhoff, The Hague, pp. 159-178.<br />

60


Kofi Agyenim Boateng<br />

Short, J., Williams, E. and Christie, B. (1976) The Social Psychology of Telecommunications, John Wiley & Sons,<br />

London.<br />

Suchman, L. (1987) Plans and Situated Actions: The Problem of Human-Machine Communication, Cambridge<br />

University Press, Cambridge.<br />

Thieme, R. B. (2003) The Plan of God, R B Thieme, Jr., Bible Ministries, Houston, Texas.<br />

Thompson, J. B. (1995) The Media and Modernity: A Social Theory of the Media, Stanford University Press,<br />

Stanford, California.<br />

Wagner, E. D. (1994), In support of a functional definition of interaction. The American Journal of Distance<br />

Education, 8, 2, 6-29<br />

Walther, J. B. (1992), Interpersonal Effects in Computer-Mediated Interaction: A Relational Perspective.<br />

Communication Research, 19, 1, 52-90<br />

Yin, R. K. (2003) Case Study Research: Design and Methods, SAGE Publications, Thousand Oaks.<br />

Yoo, Y. and Alavi, M. (2001), Media and group cohesion: Relative influences on social presence, task participation,<br />

and group consensus. MIS Quarterly, 25, 3, 371-390<br />

61


Investigating the Factors Inhibiting SMEs from Recognizing<br />

and Measuring Losses From Cyber Crime in South Africa<br />

Gino Bougaardt and Michael Kyobe<br />

University of Cape Town, Cape Town, South Africa<br />

michael.kyobe@uct.ac.za<br />

Abstract: The level of cyber attacks on organisations has increased tremendously in recent years. When such<br />

attacks occur, organisations need to assess the damage and loss from this crime. While large organisations have the<br />

mechanisms to determine such losses, SMEs lack such capability and often ignore the need to implement effective<br />

information security measures (Kyobe, 2008; Altbeker, 2000; Upfold and Sewry, 2005). Consequently, their risk<br />

exposure to cyber threats and the losses they incur from these attacks are often high (Ngo, Zhou, Chonka and Singh,<br />

2009). However, the current legislative requirements, costly legal liabilities for non-compliance, and increasing<br />

pressure by stakeholders (e.g., lenders, business partners) on SMEs to comply with good practices suggest that<br />

SMEs cannot ignore security any longer. In order to ensure accountability and compliance with security<br />

requirements, it is imperative for SMEs to identify, account and report cyber incidents and losses resulting from cyber<br />

attacks. This study investigated the factors that inhibit SMEs from recognizing and measuring losses from cyber<br />

attacks in South Africa. A survey involving twenty organisations from different business sectors was conducted and<br />

the results indicate that victimisation, resulting from a lack of awareness of cyber-crime has the greatest influence on<br />

SMEs’ ability to recognise and prepare losses from cyber attacks.<br />

Keywords: cybercrime, recognition and measuring losses, SMEs, victimisation<br />

1. Introduction<br />

Despite the availability of numerous methods and publications on how organisations can manage<br />

information security risks, SMEs still face serious challenges in managing cybercrime and resulting<br />

losses. Cybercrime are criminal activities involving the use of electronic devices and may lead to<br />

incidents such as theft of information; Sabotage of data of networks; loss of information due to<br />

eavesdropping ; financial fraud; denied access to information; and damage due to virus attacks (Kshetri,<br />

2009). These incidents result in financial and other losses to the organisations.<br />

Many SMEs fail to identify or recognise and account for these losses. They fail to manage these risks and<br />

continue to ignore the implementation of effective information security measures (Kyobe, 2008; Altbeker,<br />

2000; Upfold and Sewry, 2005). Consequently, their risk exposure to cyber attacks and the resulting<br />

losses continue to rise (Ngo, Zhou, Chonka and Singh, 2009). The current legislative environment,<br />

pressures from stakeholders (e.g., lenders, business partners) and liabilities in the event of violation of<br />

legal requirements make it imperative for SMEs to assess or estimate security risks and account for<br />

potential losses resulting from cyber attacks. Risk assessment is the process by which systems risks are<br />

identified and assessed in order to justify safeguards and protect systems from attacks (Carroll, 1996).<br />

Accounting for losses involves the evaluation and reporting of the damages when they occur. While<br />

studies on information security in SMEs are increasing, limited attention has been paid to SME inability to<br />

recognise and account for losses from cyber-crime.<br />

The objective of this study is to identify the factors that inhibit SMEs from identifying/recognizing and<br />

measuring losses from cyber attacks. In addition, we also determine the degree of influence each of the<br />

identified factors has on SME ability to recognise and prepare losses. In the following sections, the<br />

authors review literature on factors inhibiting recognition and measurement of losses. A conceptual model<br />

representing the relationship between these factors is presented. The results of an empirical test of this<br />

model are then presented. This is followed by conclusions and recommendations for future research.<br />

2. Literature review<br />

The process of identifying, recognising and reporting losses from cyber crime has not been easy for<br />

many organisations. Numerous challenges have been identified. For instance, difficulties in<br />

understanding what cyber crime represents, difficulties relating to risk identification and analysis,<br />

weaknesses in data recording and interpretation, poor design of security systems, risk management and<br />

human behaviour (Canhoto, 2010). Some of these challenges are examined in more detail below:<br />

62


Gino Bougaardt and Michael Kyobe<br />

2.1 Lack of awareness of cybercrime and lack of knowledge of being a cybercrime<br />

victim<br />

In his position paper to the Oxford Internet Institute, Baker (2010) argues that the process of collection<br />

and analysis of cybercrime data is often affected by a lack of understanding of what cybercrime means or<br />

represents. This lack of understanding of cybercrime, which is also evidenced by various ambiguous and<br />

conflicting interpretations of the term, impedes its recognition and measurement. Many organisations<br />

continue to be victimised because they are not aware of the nature or characteristics of this crime.<br />

Theories of victimisation explain the persistent victimisation of individuals and organisations by cyber<br />

criminals. Gottfredson and Hirsch (1990) in their General theory of crime, and Shreck et al (2003), in their<br />

work on victimisation show that victimisation results from low levels of self-control. A lack of self-control<br />

(defined here as the inability to control oneself or one’s emotions), is reflected in behaviours like shortsightedness,<br />

being insensitive or being impatient. SME managers are reported to be insensitive to<br />

cybercrime (Jacque, 2003; Zorz, 2003). Individuals with low self-control are believed to make decisions<br />

exclusive of those situations that increase their vulnerability and fail to change or mitigate their risk<br />

factors after the first victimisation (Forde and Kennedy (1997). This also renders several victimisation<br />

surveys inaccurate as they often underestimate incidence and prevalence rates (Fafinski, 2010).<br />

Solms and Solms (2004) assert that the lack of awareness sin is still committed by many companies.<br />

There are no proper awareness training programs and consequently, users are unaware of the risks of<br />

using their IT infrastructure and the potential damage they can cause to it.<br />

2.2 Lack of risk management skills<br />

The goal of risk management is to identify, measure, control and minimize losses associated with<br />

uncertain events (Patel and Zaveri, 2010). SMEs fail to recognise and measure losses because they do<br />

not engage in record keeping and IT risk assessment and management (Dimopoulos, Furnell and Barlow,<br />

2003). Dimopoulous et al. (2003) attributes this to the lack of funds, expertise and awareness of security<br />

risks. Risk analysis is perceived by SMEs as being complex, requiring specialist expertise and therefore<br />

something to be outsourced (Dimopouloulous et al (2003). It is also thought to disrupt management and<br />

employee activities throughout its duration. Furthermore, existing models (e.g. CRAMM) for evaluating<br />

the benefits of reducing the risks versus the investment in security technology are difficult to understand<br />

or use by SMEs (Dimopoulous et al., 2003). There are also challenges involved in comprehending the<br />

results and reports generated by these tools. The identification of risks is made harder for smaller<br />

organisations due to changes in technology and modes of vulnerabilities (Srinivasan and Abi-raad, 2003).<br />

2.3 Information system security design/infrastructure<br />

The design of the security system may also impact on the recognition and estimation of losses. Modern<br />

business environment comprises of many different applications and systems and each of these has its<br />

own threat profile (Conklin and Dietrich,2008). Such environment creates challenges for security<br />

practitioners responsible for developing security solutions. Consequently, these developers are forced to<br />

come up with piecemeal security designs, often disjointed, patched and can not monitor and<br />

comprehensively report on the security environment in the organisations (Conklin and Dietrich, 2008).<br />

Acccording to Canhoto (2010:1), the technical characteristics of the environment also influence how data<br />

on losses is derived and expressed e.g., data format, content and threshold, and which alerts and reports<br />

are produced. He states further that “formal aspects of the environment such as policies or regulations<br />

provide general definition of cybercrime, and may specify signs of alarm and the expected behaviour from<br />

analysts that detect such signs”. SMEs usually do not possess security and compliance policies (Kyobe,<br />

2008), and as indicated above, many do not engage in formal planning. Therefore such signs of alarm<br />

are often not identified before hand.<br />

2.4 Management attitude to security<br />

Individual cognitive processes, e.g., expectations, stereotypes and prior experiences, may influence<br />

attitude to security and the nature of data identified (Canhoto, 2010). Researchers argue that the failure<br />

by entrepreneurs and small business managers to proactively implement measures to handle risks has<br />

more to do with their personal characteristics (Ndubusi et al., 2005; Sjöberg et al., 2004; Nattaradol,<br />

2002; Orford et al., 2004). Proactive-risk handling is defined as the process in which potential risks to a<br />

business are identified in advance, analyzed, mitigated and prevented, and the cost of protection<br />

63


Gino Bougaardt and Michael Kyobe<br />

balanced with the cost of exposure to the risk. This does not appear to be done by small business<br />

managers or entrepreneurs. For instance, many entrepreneurs in South Africa are reported to have<br />

started businesses without giving proper consideration to economic, environmental, and cognitive<br />

limitations (Orford et al., 2004; Ladzani and Netswera, 2003). In his report on ICT adoption by rural SMEs<br />

in Thailand, Nattaradol (2002) also shows that lack of proactiveness and proper evaluation of ICT<br />

projects resulted in misjudgment or under-estimation of potential business and security risks. He<br />

identified several practices in ICT adoption which are indicative of irrational planning or behaviors (e.g.,<br />

failure to estimate project costs, use of unskilled or untrained staff to manage ICT installations, use of<br />

obsolete hardware and software and ignoring potential impact of hackers and sneakers).<br />

In addition, small business managers tend to have a strong desire for autonomy and control which could<br />

easily hamper the success of the organization (Kyobe, 2006). They tend to be rigid, traditional and<br />

usually do not draw up definitive duties or responsibilities for their subordinates. The high degree of<br />

generality resulting from undertaking heterogeneous work prevent employees from developing expertise<br />

in dealing with IT security issues (Kyobe, 2006).<br />

It is also widely reported that because of their false sense of security, small business managers are<br />

complacent about cyber-attacks and often shun good security systems and practices ( Zorz,<br />

2003;Jacque, 2003). Jensen (2004) found that security only became of much greater concern for the<br />

entrepreneurs once they had adopted e-commerce and experienced the reality of risks. Freeman (1999)<br />

explains this behaviour using Kübler-Ross’ (1969) loss model. He argues that organisations response to<br />

major environmental change is similar to individual response to loss. According to Kübler-Ross’ (1969)<br />

stage theory of loss, many who suffer loss proceed through these five stages – denial, anger, bargaining,<br />

depression, and acceptance. At the first stage, the person denies that the loss is inevitable. Freeman<br />

argues that this happens to organisations as well. Citing the work of Starbuck, Greve & Hedburg (1978),<br />

Freeman shows that managers would deny the extent of organisational crises to avoid blame.<br />

Another problem related to this is the unwillingness by managers to report cyber attack incidents. In many<br />

cases the victims withhold reporting (Lee, 1997). The importance of reporting incidents has been<br />

emphasised in many studies on information security and safety. Today, organisations are required to<br />

implement information security reporting schemes based on standards like BS-7799/ISO-17799 (ISO-<br />

IEC 27002 (ISO, 2007) (Calder and Watkins, 2005). Gonzalez (2005) maintains that information security<br />

reporting is a quality improvement process that is essential to reduce incidents and Sveen et al (2007)<br />

state that to learn from an incident and avoid it in the future the incident’s causes must be investigated by<br />

competent people. They caution however that the quality of an investigation is a function of the resources<br />

available and the workload. If for instance, resources are fewer than the allocated workload, this quality<br />

may be compromised. Phimister et al (2003) add that sporadic emphasis and management fear of liability<br />

may also hinder success in an incident reporting system. Eurim (2003) outlines several barriers to<br />

reporting including concern about confidentiality, disruption to business and loss of reputation. A culture<br />

of reporting does not exist in many SMEs due similar reasons outlined above (Kyobe, 2004;2006).<br />

2.5 Lack of knowledge of, and compliance with security regulations<br />

Cybercrime regulations require that organisations and individual recognise and account for damages<br />

resulting from cybercrime. Therefore a lack of knowledge of these security regulations and compliance<br />

with their requirements suggest inability to recognise and report on cyber-crime risks and losses. In South<br />

Africa cybercrime is regulated by the cyber crime section in Chapter XIII of the ECT Act, 2002<br />

(Michalson, 2005). This chapter introduces statutory criminal offenses relating to unauthorized access to<br />

data (e.g., through hacking), interception of data (e.g., tapping into data flows or denial of service<br />

attacks), interference with data (e.g., viruses) and computer related extortions, fraud and forgery. They<br />

also state that a person aiding those involved in these crimes will be guilty as an accessory. A person<br />

convicted of an offence related to the above is liable to a fine or imprisonment for a period not exceeding<br />

five years.<br />

Kyobe (2008) found that majority of the SMEs surveyed were not proactive in their approaches to<br />

compliance with such regulations. Few firms planned their security systems, allocated sufficient<br />

resources for compliance, trained their staff regularly and developed policies. Economical factors such as<br />

indirect compliance costs were found to be the main barriers to SME compliance in this study. Kyobe<br />

(2008) found that of the 30 South African Websites investigated, 40 percent had not implemented<br />

necessary legal requirements. 54% of all respondents never reported their effort towards compliance to<br />

64


Gino Bougaardt and Michael Kyobe<br />

auditors and 32 percent of the respondents were not aware of the requirements and liabilities of the ECT<br />

Act.<br />

2.6 Summary<br />

The key relationships as described in the previous sections are presented in the conceptual model in<br />

Figure 1 below. This model was then validated empirically in a study involving 20 SMEs as indicated in<br />

the following sections.<br />

Quality of Information system<br />

security design/ infrastructure<br />

Expertise in Info. Security & risk<br />

management<br />

Management attitude to security<br />

Awareness of cyber-crime and of<br />

being a victim of such crime<br />

Awareness of (and compliance with)<br />

security Regulations<br />

Figure 1: Conceptual model<br />

3. Methodology<br />

Recognition of Cyber crime<br />

and preparation of losses from<br />

this crime<br />

The sample consisted of SMEs in South Africa engaged in e-commerce activity since these are expected<br />

to experience greater risk exposed to e-crimes. SMEs across South Africa from various provinces and<br />

industries were targeted in order to generalise the findings (Saunders, et al., 2009, p. 217). The<br />

participants included senior management and employees responsible for IT infrastructure. These were<br />

specifically selected because they would be involved in making decisions relating to IT and its<br />

implementation. It would also enable us to collect and compare the views of different categories of<br />

respondents. An up-to-date database from the South African SME Toolkit web site<br />

(http://southafrica.smetoolkit.org) was used to identify the SMEs. The content administrator of this web<br />

site confirmed to us that the majority of the businesses registered on the web site are SMEs and that their<br />

details are updated annually. In order to increase the sample size, more SMEs were obtained from other<br />

sources, e.g., Waverly Business Park, Business Broadcaster, Facebook and postgraduate students.<br />

Twenty two responses were unfortunately received in total. This is however a typical problem in many<br />

SME surveys where responses rates are known to be very low (Kyobe, 2008).<br />

4. Data collection<br />

A questionnaire was used to collect the data since many of the organisations involved are located in<br />

different areas. The questionnaire was piloted with academics in the department. After making necessary<br />

65


Gino Bougaardt and Michael Kyobe<br />

adjustments, this was then emailed to the respondents together with a cover letter explaining the purpose<br />

of the study. The constructs and their measures were mainly adapted from previous research papers<br />

(Kyobe,2008; Upfold and Sewry, 2005; and Herath and Rao, 2009).<br />

Table 1: Constructs and measures<br />

Construct Measure Questions<br />

Adapted from<br />

Quality of Information Security<br />

Design/Infrastructure<br />

Lack of expertise in (Information<br />

security & risk management)<br />

-Our information security solution protects the entire<br />

business system<br />

-Use up-to-date software and Hardware<br />

-Use anti-virus software<br />

-Use a firewall and data encryption<br />

-Availability of IS/IT expertise in security<br />

-Seek expert assistance from external sources<br />

-Techniques employed are useful in determining<br />

cyber losses<br />

-experience difficulties using estimation techniques<br />

Management attitude to security -Mitigate threats or risks to business/IT systems<br />

-Discipline employees who violate security<br />

regulations<br />

- Train staff on IS risks<br />

- false sense of security<br />

Lack of awareness of cyber-crime<br />

and of being a victim of such<br />

attacks<br />

Awareness and compliance with<br />

regulations<br />

Recognise computer crime and<br />

prepare losses<br />

- readness/willingness to report incidents<br />

-Our system is often unavailable due to cyber attacks<br />

-Have suffered financial loss due to cyber attacks<br />

-<br />

- have documented a Security policy<br />

-Comply with the ECT Act<br />

-employees are aware of the ECT Act requirements<br />

and penalties<br />

- Perform record keeping<br />

-availability of expertise in preparing loss estimates<br />

- prepare financial loss estimates due to computer<br />

attacks<br />

-possess expertise in preparing financial losses<br />

we document information security activities<br />

- we conduct information security audits<br />

Kyobe (2008)<br />

Upfold and Sewry<br />

(2005)<br />

Kyobe(2006,<br />

2008)<br />

Kyobe(2008)<br />

Kyobe(2008)<br />

Upfold and Sewry,<br />

2005; and Herath<br />

and Rao, 2009<br />

The first part of the questionnaire obtained general information about the respondents. Subsequent<br />

sections measured the different aspects of the research model (i.e., quality of the security infrastructure,<br />

expertise in information security management, the attitudes of management towards information security,<br />

awareness of cyber-crime and acknowledgement of being a victim of such crime, knowledge of the<br />

security regulation and the extent to which SMEs recognise cyber crime and prepare losses..The<br />

demographic information of the SMEs is presented in Table 2 below.<br />

Table 2 indicates that 9 of the organisations are from the ICT business sectors. There are 11<br />

organisations based in the Western Cape and 6 organisations based in Gauteng. Further analysis<br />

revealed that 8 of the respondents were IT/IS staff while the rest were business owners/managers).<br />

Furthermore, business managers had on average internet trading experience of 5 years and computer<br />

experience of 15 years while IS/IT staff had 4 years and 11 years respectively.<br />

5. Reliability assessment<br />

The reliability test was conducted using the Cronbach’s alpha (Saunders, et al., 2009, p. 374). The<br />

Cronbach alpha coefficient above the recommended 0.70 would confirm the reliability of the questions<br />

66


Gino Bougaardt and Michael Kyobe<br />

used (Nunnally, 1978). Table 3 shows the reliability test results for the constructs in the conceptual<br />

model.<br />

Table 2: SME demographics<br />

1 4<br />

No. of<br />

Employee<br />

s Business Sector Location<br />

2 30<br />

Yrs<br />

trading<br />

on<br />

Internet<br />

Respondent<br />

Position<br />

Respondent<br />

yrs of<br />

computer<br />

experience<br />

Engineering<br />

Consulting<br />

Information<br />

Western Cape Owner<br />

Technology Western Cape Consultant (IT) 15<br />

3 12 NGO<br />

Broadband<br />

National 0 Software Developer 10<br />

4 150 provider-ICT<br />

Procurement to<br />

Western Cape 11 BP Manager- 9<br />

5 9 Mining & Industry<br />

Information<br />

Gauteng 4 Manager 10<br />

6 1 Technology Western Cape 1 Owner 20<br />

7 6 Telecoms North West 10 Sales Director 18<br />

8 2 Accounting Western Cape<br />

Limpopo/KZN/M<br />

4 Director 11<br />

9 24 Construction pumalanga Director 10<br />

10 2 IT GT 5 Director 15<br />

11 15 IT Consulting Gauteng 3 CEO<br />

CEO/Project<br />

17<br />

12 1 ICT Gauteng 1<br />

Manager 4<br />

13 5<br />

Management<br />

Consulting Gauteng 2 Managing Director 10<br />

14 1 Consultancy<br />

Information<br />

Western Cape 2 Owner 27<br />

15 7 Technology KZN 2 MD 20<br />

16 8<br />

Financial<br />

Services Western Cape 11 Director 20<br />

17 46 Building Western Cape 6 CEO 12<br />

18 (1)<br />

Financial<br />

Services Western Cape<br />

Western<br />

10<br />

19 80 Market Research Province 0<br />

Senior Architect/<br />

Strategist 16<br />

IT Development<br />

Team Leader 10<br />

20 80 Online Gaming Western Cape 5 Architect 5<br />

21 47 Construction Western Cape 9 Marketing Director 11<br />

22 20 Insurance Gauteng 3 Director 25<br />

The Cronbach alpha coefficients (in Table 3) are above the recommended 0.70 (Numally, 1978), thereby<br />

confirming reliability or internal consistency of the variables used in the present study. Most variables as<br />

stated above were adapted from prior studies. This also assisted in ensuring construct and content<br />

validity.<br />

67


Table 3: Assessment of reliability<br />

6. Analysis<br />

Gino Bougaardt and Michael Kyobe<br />

Constructs No.of<br />

Items<br />

Cronbach<br />

Alpha<br />

Quality of IS Security Design 3 0.70<br />

Lack of Expertise: Info. Security 4 0.82<br />

Management attitude to security 4 0.73<br />

Awareness of cybercrime (victimisation) 4 0.88<br />

Awareness of (and compliance with)security regulations 3 0.75<br />

Recognise computer crime & prepare losses from cyber crime 5 0.82<br />

Table 4: Mean responses<br />

No. of All respondents Business IT/IS staff<br />

Items<br />

Managers<br />

Constructs Mean Std Mean Std Mean Std<br />

Quality of IS Security Design 3 3.41 1.13 2.99 0.65 3.77 0.77<br />

Lack of Expertise: Info. Security 4 2.99 0.99 3.50 1.00 3.50 0.99<br />

Management attitude to security 4 2.49 0.91 2.19 0.65 3.90 0.97<br />

Awareness of cybercrime (victimisation) 4 3.01 1.16 2.27 0.88 3.98 0.88<br />

Awareness/compliance with security<br />

regulations<br />

3 2.45 0.73 2.59. 0.66 3.31 1.12<br />

Recognise & prepare losses from cyber crime 5 2.47 0.37 2.84 0.60 3.67 0.76<br />

6.1 Correlation analysis<br />

Correlation analysis was conducted to determine the association between variables.<br />

Table 5: Correlation analysis<br />

1 2 3 4 5 6<br />

1. Quality of IS security design 1<br />

2. Lack of expertise in IS 0.21 1<br />

3. Management attitude to security 0.39 0.44 1<br />

4. Awareness of cyber crime attacks (victimisation) 0.11 0.22 0.35 1<br />

5. Awareness/compliance with regulations 0.31 0.33 0.07 0.48 1<br />

6. Recognise & prepare losses from cyber crime 0.18 0.28 0.21 0.37 0.19 1<br />

* (correlations significant at .05 or less are shown in bold)<br />

6.2 Regression analysis<br />

Results of the regression analysis (Table 6) suggest that in the South African SMEs, lack of awareness or<br />

victimisation has the most influence on SMEs’ ability to recognise and measure losses from cyber crime<br />

than lack of expertise in information security management and management attitude to security. It is<br />

68


Gino Bougaardt and Michael Kyobe<br />

surprising to find in this study that quality of IS design and awareness and compliance with regulations<br />

did not have significant influence on recognition and measuring losses.<br />

Table 6: Regression summary for dependent variable: Recognising & measuring losses from cyber crime<br />

R = 0.8209 ; R Sq = 0.6739; Adjusted R Sq = 0.5720; F(5,16) =6.6132; p < 0.00162; Std.Error of estimate:<br />

0.68841<br />

Beta Std Err of<br />

Beta<br />

B Std Err of B T(16) P=level<br />

Intercept 2.5642 1.8677 1.8436 0.01887<br />

Quality of IS design -0.1238 0.1611 -0.2151 0.2798 -0.7678 0.45232<br />

Lack of expertise 0.3394 0.1895 0.35110 0.1959 1.7914 0.04932<br />

Management attitude<br />

to security<br />

0.4621 0.2177 0.1299 0.0612 1.1234 0.04365<br />

Lack of awareness<br />

(victimisation)<br />

0.8119 0.2034 1.11080 0.2783 3.990 0.00105<br />

Awareness &<br />

compliance with<br />

regulations<br />

0.5262 0.2431 1.0391 0.4800 2.1648 0.4589<br />

7. Discussion of findings<br />

Table 4 shows that most respondents did not consider the quality of IS security design to be good. The<br />

mean score on this item was 3.41 (uncertain). Respondents were not certain whether the security of the<br />

system (e.g., Firewall, software) really protected their businesses. Almost similar responses were<br />

obtained for the level of security expertise (2.99). Table 2 indicates that ½ of the respondents had more<br />

than 10 years of IS/IT experience. This suggests that possession of IT skills does not necessarily<br />

translate into proper management of IS security risks. There was a positive and significant association<br />

between information security design/infrastructure and the dependent variable (see Table 5, item 6 -<br />

recognise and prepare losses from cyber crime).Table 5 shows a positive and significant association<br />

between quality of information security design and the dependent variable. This suggests therefore that<br />

the former does influence the latter. Insufficient awareness of IT risks and computing limitations are major<br />

factors inhibiting small organizations from engaging in effective planning and monitoring of business<br />

operations (Kyobe, 2004).<br />

Table 4 also shows that overall most respondents seem to ignore information security requirements. The<br />

mean score was 2.49 indicating rather a negative attitude or lack of attention to security requirements.<br />

Management attitude to security was measured by asking respondents to indicate whether they mitigated<br />

risks, disciplined employees who violated security requirements, trained staff on IS risks and reported<br />

security violations. Further analysis of the responses of business managers and those in IT/IS staff<br />

confirms that information security is not a major concern for business managers (mean score for<br />

business managers was 2.19 compared to 3.90 by IT/IS staff). It is not surprising therefore that Table 5<br />

reveals a significant correlation between Management attitude to security and (quality of IS security<br />

design and lack of expertise in IS security). These influence SME ability to recognise losses as confirmed<br />

by the positive and significant relationship revealed in Table 5.<br />

In the case of awareness of cybercrime (and victimisation), the overall mean response was 3.01 (see<br />

Table 4), which suggests the respondents were uncertain. However, a breakdown of the responses<br />

indicates that only those with technical skills appear to be aware that their systems were unavailable due<br />

to cyber attacks and that they do suffer losses due to cyber attacks (mean score was 3.98). Business<br />

managers were mainly uncertain of the potential for victimisation.<br />

Compliance with security regulations remains a major problem in these organisations. The mean score<br />

was 2.45 (disagree). This may not be surprising given the fact that most business managers were also<br />

found to be least concerned about information security risks and also did not know they were potential<br />

victims of cyber attacks. A further analysis show that even IT/IS staff were not certain whether they<br />

complied with security regulations (mean score was 3.31). This confirms Jacque’s (2003) observation<br />

that SME managers were insensitive to cybercrime. Such level of insensitivity translates into lack of self<br />

control which results into re-victimisation (Shreck, 2003). The fact that this construct was found to be<br />

associated with recognition and reporting of losses from cyber-attacks confirms that lack of compliance<br />

influence SME ability to prepare losses.<br />

69


Gino Bougaardt and Michael Kyobe<br />

Most respondents did not agree that they do recognise and prepare losses from cyber attacks (Mean<br />

score 2.47). However, some IT/IS staff seemed to agree that they do to some extent (3.67). Recognition<br />

and preparation of losses from cyber attacks was measured by asking respondents to indicate if they<br />

kept records, possessed expertise in preparing losses, document information security incidents and<br />

conduct security audits (see Table 1). Table 5 shows that all other constructs were significantly correlated<br />

with this construct.<br />

However, when these relationships are analysed in a regression analysis, it is surprising to find that the<br />

quality of IS design, and awareness and compliance with regulations have no significant influence on<br />

recognition and measuring losses. This could perhaps be explained by the small sample size used in the<br />

present study<br />

Results of the regression analysis (Table 6) suggest that in these South African SMEs, lack of awareness<br />

of cyber crime (or victimisation) has the most influence on SMEs’ ability to recognise and measure losses<br />

from cyber crime than lack of expertise in information security management and management attitude to<br />

security. It is surprising to find in this study that quality of IS design and awareness and compliance with<br />

regulations did not have significant influence on recognition and measuring losses. This could also be<br />

attributed to the small sample size and the number of items used to measure these constructs. It is<br />

however interesting to find that our suspicion regarding three of the constructs (i.e., lack of awareness of<br />

cyber-crime (victimisation); management attitude to security; and lack of expertise in risk management)<br />

were found to be correct in this analysis.<br />

8. Conclusion<br />

This study shed more light on those factors influencing preparation of losses from cyber attacks in SMEs.<br />

Lack of awareness of cyber-crime appears to be a serious challenge for many SMEs and it is not<br />

surprising that these organisations continue to be victimised. Because SMEs managers do not<br />

understand what cyber-crime is, they simply may not take it seriously and may not find it necessary to<br />

engage in risk management initiatives. Risk management systems are vital to the evolution of SMEs<br />

(Sanchez, Ruiz, Fernandez-Medina, and Piattini, 2010). Management need to engage in risk<br />

management practices but this requires a change in attitude or in the way information security is<br />

perceived. Upfold and Sewry (2005) reported that SME management view information security policies as<br />

hard and expensive. SMEs managers can not continue rejecting good security practices due to their false<br />

sense of security (Kyobe, 2008).<br />

The findings presented in the present paper are however based on evidence gathered from only twenty<br />

two SMEs. Therefore precautions need to be taken when generalizing these findings. This study should<br />

be repeated with a much larger sample and the relationships between the constructs tested again in a<br />

regression analysis. Given the limited sample size, we could not compare the responses of business<br />

managers and IT/IS staff. This could reveal more details about the potential causes of different behaviors<br />

of managers towards security. Future studies should investigate these relationships and their impact on<br />

the dependent variable.<br />

References<br />

Altbeker, A. (2000). E-security: The race against computer crime. Retrieved March 06, 2010, from<br />

http://www.iss.org.za/PUBS/CRIMEINDEX/00VOL4NO6/ESecurity.html<br />

Baker, W.H. (2010). Thoughts on Mapping and Measuring Cybercrime.<br />

Oxford Internet Institute Forum Mapping and Measuring Cybercrime. http://www.sfu.ca/~icrc/content/oxford.foru<br />

m.cybercrime.pdf. Page 26.<br />

Calder, A., and Watkins, S., (2005), IT Governance, 3rd edition, Kogan Page, London<br />

Canhoto, A. (2010). What’ before ‘How’. Oxford Internet Institute Forum .<br />

http://www.sfu.ca/~icrc/content/oxford.forum.cybercrime.pdf.<br />

Carroll, J. M. (1996). Computer Security. Third Edition, Newton: Butterworth- Heinemann<br />

Conklin, W. A., and Dietrich, G. (2008, January 7). Systems Theory Model for Information Security. Proceedings of<br />

the 41st Hawaii International <strong>Conference</strong> on System Sciences - 2008, (pp. 1-9). Big Island, Hawaii.<br />

Dimopoulos, V., Furnell, s.m., and Barlow, I.M. (2003).Considering IT Risk Analysis in Small and Medium<br />

Enterprises. Proceedings of the 1st Australian Information Security Management <strong>Conference</strong> 2003 (InfoSec03),<br />

Perth, Australia, 24 November 2003<br />

Available: scissec.scis.ecu.edu.au/proceedings/2003/infosec/pdf/02_final.pdf<br />

EURIM (2003) IPPR E-Crime Study Partnership Policing for the Information Society. Working Paper 1: Reporting<br />

Methods and Structures. http://www.eurim.org/consult/e-crime/dec03/ECS_WP1_web_031209.pdf<br />

Fafinski, S. (2010) Mapping and Measuring Cybercrime, Position Paper. p75.<br />

Mapping and Measuring Cybercrime. http://www.sfu.ca/~icrc/content/oxford.forum.cybercrime.pdf. Page 36<br />

70


Gino Bougaardt and Michael Kyobe<br />

Forde, D., and Kennedy, L. (1997). Risky lifestyles, routine activities, and the general<br />

theory of crime. Justice Quarterly, 14, 265-289.<br />

Freeman, S. (1999). Identity Maintenance and Adaptation: A Multilevel Analysis of Response to Loss. Research in<br />

Organizational Behavior, 21, 247-294.<br />

Gonzalez, J.J. (2005), Towards a Cyber Security Reporting System – A Quality Improvement Process, Springer,<br />

Berlin Heidelberg<br />

Gottfredson, Michael R., and Travis Hirschi. 1990. A General Theory of Crime. Stanford, CA: Stanford<br />

University Press.<br />

Herath, T., and Rao, H. R. (2009). Encouraging information security behaviors in organizations: Role of penalties,<br />

pressures and perceived effectiveness. Decision Support Systems , 47 (2), 154-165.<br />

Jaques, R. (2003). Survey finds firewall and antivirus software considered unimportant by SMEs, URL (Consulted<br />

September, 2004): http://www.frame4.com/php/printout689.html<br />

Kyobe, M. (2006). Entrepreneur behaviors on e-commerce security, In M. Khosrow-Pour (ed.) Encyclopedia of ecommerce,<br />

e-government, and mobile commerce, pp. 437-444).<br />

Kyobe, M. (2008). Evaluating Information Security within SMEs engaged in E-commerce in South Africa. Retrieved<br />

March 08, 2010, from http://www.isbe.org.uk/Kyobe<br />

Ladzani, M.W., and Netswera, F.G. (2003). Lessons from successful South African rural entrepreneurs, URL<br />

(Consulted October, 2007): http://www.sbaer.uca.edu/research/icsb/2003/papers/48.doc<br />

Lee, W. (1997). A Deterrent Measure Against Computer Crime:<br />

Knowledge-Based Risk-Analytic Audit, Singapore Management Review,<br />

January, pp 19-45.<br />

Michalson, (2009). “ Protection of Personal Information Bill - the implications for you”. Michalson.com. [Online]<br />

Available: http://www.michalsons.com/protection-of-personal-information-bill-the-implications-for-you/3041.<br />

Access date: 10 January 2010<br />

Nattaradol, P. (2002), Harnessing ICT potential for the benefits of farmers and the rural poor: Experience and vision<br />

of Bank for agricultural cooperatives (BAAC), Thailand. BAAC. URL (Consulted October, 2007):<br />

http://www.adb.org/annualmeeting/2002/seminars/presentation/nattaradol_presentation.pdf.<br />

Ndubisi, N.O., Gupta, O.K., and Ndubisi G.C. (2005). The Moguls’ Model of Computing: Integrating the Moderating<br />

Impact of Users’ Persona into The Technology Acceptance Model, Journal of Global Information Technology<br />

Management, 8(1): 27-47.<br />

Ngo, L., Zhou, W., Chonka, A., and Singh, J. (2009). Assessing the Level of I.T. Security Culture Improvement:<br />

Results from Three Australian SMEs. IECON '09 35th Annual <strong>Conference</strong> of IEEE 2009, (pp. 3189-3195).<br />

Porto, Portugal.<br />

Nunnally, J. C. (1978). Psychometric Theory. New York: McGraw-Hill.<br />

Orford, J., Herrington, M., and Wood, E. (2004). South African Report Global Entrepreneurship Monitor. Retrieved<br />

January 22, 2005, from www.gsb.uct.ac.za/gsbwebb/userfiles/GEM_ 2004.pdf .<br />

Patel, S. and Zaveri, J. (2010). Assessment model of cyber-attacks on information systems. Journal of computers,<br />

15(3): 352-359.<br />

Phimister,J., Oktem,U., Kleindorfer,P.R, and Kunreuther, H. (2003). Near-Miss Incident Management in the<br />

Chemical Process Industry. Risk Analysis, Volume 23, Issue 3, pages 445–459, June 2003<br />

Sanchez, C.R, Fernández-Medina,E., and Piattini, M. (2010): Managing the Asset Risk of SMEs. ARES 2010: 422-<br />

429<br />

Saunders, M., Lewis, P., and Thornhill, A. (2009). Research methods for business students (5th ed.). Harlow:<br />

Pearson Education Limited.<br />

Schreck, C.J., and Miller, J.M. (2003). Sources of fear of crime at school: What is the relative contribution of<br />

disorder, individual characteristics, and school security?, Journal of School Violence 2 (2003), pp. 57–77.<br />

Solms, B, and Solmns R. (2004). The 10 deadly sins of information security management, Computer & Security<br />

23(5): 371-376.<br />

Sjöberg, L., Moen,B., and Rundmo, T. (2004). Rotunde Publications 84. Explaining risk perception. An evaluation of<br />

the psychometric paradigm in risk perception research, URL (Consulted February, 2005):<br />

http://www.svt.ntnu.no/psy/Torbjorn.Rundmo/psychometric_paradigm.pdf<br />

Srinivasan, G., and Abi-raad, M. (2003). Risk factors associated with e-buziness infrastructure of SMEs. Paper 21. 1 st<br />

Australian Information Security Management <strong>Conference</strong>, Perth, W.Australia.<br />

Sveen, F., Rich, E., and Jager, M. (2007), "Overcoming organizational challenges to secure knowledge<br />

management", Information Systems Frontiers, Vol. 9 No.5, pp.481.<br />

Upfold, C. T., and Sewry, D. A. (2005). An Investigation Of Information Security In Small And Medium Enterprises<br />

(SME's) In The Eastern Cape. Retrieved March 06, 2010, from<br />

http://icsa.cs.up.ac.za/issa/2005/Proceedings/Research/082_Article.pdf<br />

Zorz, J.(2003, May 15). Small firms ‘shun’ PC security. BBC NEWS. Retrieved October 15, 2004, from<br />

http://www.net_security.org/news:php?id=2650.<br />

71


Sharing knowledge – The CoP way<br />

Sheryl Buckley and Apostolos Giannakopoulos<br />

University of Johannesburg, South Africa<br />

sbuckley@uj.ac.za<br />

paulg@uj.ac.za<br />

Abstract: Managing academics’ knowledge to the benefit of all is a challenge to any institution. One way to share<br />

knowledge is through a community of practice. However, knowledge sharing within communities is complex.<br />

Challenges such as a lack of trust, a lack of incentives and a poor culture of learning in an institution have an effect<br />

on effective knowledge sharing. A questionnaire consisting of open- and closed-ended questions was administered<br />

to academics. This article reports on academics’ awareness and views of a community of practice at a higher<br />

education institution in South Africa. Empirical evidence shows that academics would be willing to share knowledge.<br />

However, time constraints, an unwillingness among academics to share knowledge, and a lack of support or<br />

participation from management could become obstacles to knowledge sharing.<br />

Keywords: knowledge sharing, communities of practice, higher education, trust, learning culture, knowledge<br />

1. Introduction<br />

Knowledge is the driving force of a knowledge economy. Therefore, the way knowledge is shared and<br />

created and the way these actions are managed could lead to either a competitive advantage for an<br />

organisation or it could lead to its demise. The same rationale applies to an academic institution. The<br />

knowledge that academics possess, forms the largest percentage of such an institution’s market value.<br />

Since knowledge (both tacit and explicit) resides in the minds of the people and some of it can be<br />

subsequently codified and become ‘common to all knowledge’ (Smith and McKeen 2004, 398), managing<br />

academics’ knowledge becomes a challenge to the institution.<br />

One way to share knowledge is through a community of practice (CoP) (Wenger 2004). A CoP consists<br />

of groups of people who come together voluntarily, both face-to-face and virtually, to share and to learn<br />

from one another. For a CoP to exist there must be people, shared interest, shared competencies and<br />

shared activities performed on a regular basis that advance learning and knowledge (Wenger, 2006).<br />

Knowledge sharing within communities is complex due to the lack of shared consensual knowledge or<br />

shared sense of identity (Barab and Duffy 1998; Thorpe 2003). Firstly, the development of trust is seen to<br />

be a key foundation which requires to be developed before effective knowledge sharing can occur. To<br />

facilitate effective community knowledge sharing, effort needs to be invested in developing the social<br />

relationship (and hence trust) (Muller 2006) between members of the communities. Secondly, appraisal<br />

systems need to be coupled to a proper system of rewards for those staff that enhance the institution’s<br />

stock of knowledge and use it to creative ends (Sallis and Jones 2002, 43). Lastly, a learning culture is a<br />

key component for ensuring that an institution can benefit from the knowledge at its disposal (Sallis and<br />

Jones 2002, 78). It can be argued that academics looking for knowledge connect with those who possess<br />

it and in this way they are able to improve their practice and create a competitive sustainable advantage<br />

for themselves and the institution. If the above issues are not addressed, knowledge sharing will be<br />

compromised.<br />

This article reports on participants’ awareness and views of communities of practice at a HE institution<br />

and proposes a framework for a CoP in an academic environment.<br />

2. Impact of knowledge and information sharing in higher education<br />

A university creates, investigates, evaluates and conveys culture through its scientific research and<br />

teaching activity (Brzezinski 1997, 202). Research universities are crucial factors in the development of<br />

new knowledge. Jarvis (2007, 90) writes: the university as whole, not just individual faculty members<br />

within the institution needs to be measured on the extent to which it serves the knowledge needs of<br />

society as a whole, as well as how it serves basic teaching and research.<br />

University expansion at the end of the 20 th century has led to a burgeoning of new programmes, modules<br />

and courses (Lea 2005, 180-181). The need for creating learning or knowledge communities rather than<br />

teaching factories is greater than ever before, if universities are to transform to modern universities (Msila<br />

2006, 90). As a university’s core business is teaching and learning, knowledge plays the central role in<br />

both (Jacob 2000, 141-142). The impact of failure to transmit knowledge successfully has a detrimental<br />

effect on the institution as well as the learner. Wick (2000, 519) contends that knowledge itself has limited<br />

72


Sheryl Buckley and Apostolos Giannakopoulos<br />

value because it becomes obsolete very quickly; and unless the institution ensures that its academics<br />

keep updating their knowledge, stagnation will prevail. However, it is not knowledge itself that is so<br />

valuable; rather it is the ability of an organisation’s members to generate knowledge and innovate, using<br />

that knowledge (Wick 2000, 520).<br />

For Duderstadt (2005, 81-82) “Knowledge can be created, absorbed and applied only by the educated<br />

mind, and … universities in particular, will play increasingly important roles as our societies enter this new<br />

age.” In a sense, knowledge is the medium of the university.<br />

There is widespread belief that information sharing is beneficial because it fosters collaboration, enables<br />

coordination, enhances capabilities and facilitates growth. According to (Lamberton, 2005, 155)<br />

information and communication technologies (ICT) can be used effectively as part of a toolbox for<br />

addressing global problems. Technology is a critical enabler and can be separated into two categories:<br />

infrastructure and collaborative applications (2004, 71). Infrastructure is the hardware and software that<br />

allow anyone to communicate with anyone else from any place at any time. Once the infrastructure is in<br />

place, collaborative technology (or groupware) applications can be effectively implemented to address<br />

communication across time and space.<br />

In discussing excellence in university teaching, Sherman, Armistead, Fowler, Barksdale and Reif (1987,<br />

72) consider excellence as developmental and such development as ‘a function of an interaction between<br />

individuals and their environments in an appropriate and facilitative learning atmosphere and not merely a<br />

function of the passage of time’. Such individuals are known to operate collaboratively with other<br />

colleagues as they see themselves as lifelong apprentices, always willing to improve their teaching<br />

practices. By implication, if a university aims for excellence, it is not sufficient to have highly qualified and<br />

knowledgeable academics, but such academics should interact with others to improve their practices. In<br />

an indirect way, they form unofficial CoPs, with excellence as one of their main objectives.<br />

Innovation is vital for any university wishing to be proactive in the marketplace and have the competitive<br />

edge (Du Toit 2000, 188; Skyrme 2003). Du Toit (2000, 188) argues that nowadays competitiveness in<br />

the marketplace is essential for the survival of South African universities and that the ability to sustain<br />

innovation is recognised as a strategic advantage; it has become evident that knowledge must be<br />

generated and integrated within the university at an accelerated pace. Such knowledge is essential for<br />

innovation or knowledge creation within a university. As academics are surrounded by other academics<br />

with the same purpose, it stands to reason that there should be close cooperation between them. Wenger<br />

(2004) argues that the social aspect then plays a central role in achieving their purpose.<br />

3. Challenges for institutions in implementing a community of practice<br />

Witt, McDermott, Peters and Stone (2007) believe that the education sector is now exploring the<br />

usefulness of building CoPs and connecting individuals with a common interest where knowledge is<br />

created, sustained and transformed. Already familiar to many academics is the idea of working in groups<br />

such as project teams, programme teams, special interest groups and forums to discuss, share and<br />

develop practice. According to Wenger (2004) the salient features of a CoP are the domain, the<br />

community and the practice. The domain is the area of interest, the community is formed by the<br />

relationships (conversations and discussions) between members and the practice is what community<br />

members do with learning derived from their interaction.<br />

3.1 Building trust<br />

Without trust and norms of openness, employees are likely to be concerned about the repercussions<br />

caused by expressing their views openly. When there are few norms of sharing and reciprocity,<br />

employees see little need and value in sharing knowledge with others. Employees who do not identify<br />

with the institution are unlikely to be willing to put effort into participating in knowledge-sharing activities<br />

(Pee and Kankanhalli 2007, 134). Trust is a crucial knowledge issue and it is a pivotal and essential<br />

element in long-term social relationships. However, building it is difficult and challenging at HE<br />

institutions. One of the best ways for HE leaders to engender trust is to be trustworthy themselves, and to<br />

act with integrity. Trust grows out of acts of trustworthiness (Sallis and Jones 2002, 35; Hislop 2004, 41).<br />

An HEI’s trust-building process has a number of linked stages. The first stage of building collaboration<br />

has its basis in the need to develop rapport. Groups, meetings and processes need to be established so<br />

that each employee can learn from the skills and expertise of others. It is a process of bottom-down<br />

73


Sheryl Buckley and Apostolos Giannakopoulos<br />

sharing, with management encouraging and establishing the space for teamworking to develop (Sallis<br />

and Jones 2002, 36).<br />

During the second stage the members of groups become more focused and active. This is when the<br />

prevailing corporate culture can encourage group activities by rewarding innovative thoughts and action<br />

(Chetley and Vincent 2003). At this time management needs to encourage people to look outwards<br />

towards their colleagues, rather than inwards. Network leaders need to break down old habits and ensure<br />

that the trust required for collaboration can thrive (Sallis and Jones 2002, 36). Techniques such as<br />

storytelling and learning conversations play a vital role in this process and should encourage employers<br />

to shift from hoarding to sharing knowledge. Technological infrastructures can also be useful for<br />

encouraging active collaboration (Hayes and Walsham 2000, 61; Sallis and Jones 2002, 36).<br />

To reach the third stage in the trust-building exercise, academics need to have easy access to<br />

information, ideas and solutions and need to know where old ideas can be unearthed and used again.<br />

Chat rooms, online communities and other inventive ways can be instrumental in supporting the sharing<br />

activities that build trust, and move the culture to one of knowledge sharing. As more and more people go<br />

online, informal communities spring up. Such communities allow important learning to take place; they<br />

know that learning works best where there is trust (Sallis and Jones 2002, 37).<br />

The final stage is rewarding the contribution of great teamwork. People are inclined to want some sort of<br />

scorecard that acknowledges their knowledge-sharing activity. The rewards should be personal (Sallis<br />

and Jones 2002, 37).<br />

3.2 Rewarding knowledge creation<br />

HE institutions need to recognise the value of their intellectual capital and the importance of the<br />

knowledge that their staff have. Appraisal and other feedback systems need to provide clarity about what<br />

is expected of staff, and to concentrate on staff contributions to knowledge creation. Davidson and Voss<br />

(2002, 99) emphasise that ‘[a] KM initiative will not get very far if – however actively you champion<br />

knowledge sharing – the existing remuneration system rewards knowledge hoarding’. Galunic and Weeks<br />

(1999, 189) suggest a variety of recognition and financial awards for knowledge sharing to the practice as<br />

a whole, including promoting developing intellectual capital as one of the four criteria used when<br />

determining promotion and bonuses. Liebowitz and Chen (2003, 431) feel that knowledge sharing is its<br />

own reward and that introducing formal incentive schemes might have the opposite of their intended<br />

effect. Among the incentives named by Liebowitz and Chen (2003, 432) are recognition; duty or need; a<br />

good frame of reference; a sense of give and take (quid pro quo); feedback mechanisms for letting<br />

knowledge sharers know their knowledge is being put to use; and the pleasure of helping someone attain<br />

their goals.<br />

HE institutions need to show their appreciation of staff contributions to the institution’s intellectual capital.<br />

Measurement could include skills acquisition as well as a willingness to undertake additional training and<br />

research output (Sallis and Jones 2002, 43).<br />

3.3 Culture of learning<br />

HE institutions need to build a culture of learning that allows for continuous knowledge creation and<br />

transformation (Sallis and Jones 2002, 77). Argyris (1994, 82) argues that institutions fail when their<br />

culture inhibits learning. A learning organisation encourages its staff to learn by giving them the<br />

mechanisms to share experience and best practice, and to improve their skills and capabilities. Leonard<br />

(1995, 63) identified a number of important implications for educational institutions. First, the importance<br />

of what she calls ‘firm specific skills’ or tacit knowledge indicates why it is difficult to replicate the<br />

performance of a successful institution. She argues that innovations cannot simply be imported from one<br />

institution to another. Educational institutions need to focus on developing the organisational-specific<br />

knowledge of its members; talent needs to be home-grown.<br />

Learning organisations have a knowledge-sharing culture, with knowledge as the ‘cement’ holding them<br />

together. The development of informal knowledge communities alongside more formal structures<br />

empowers staff to learn as they work, and builds a common language and context in which tacit<br />

knowledge can be shared effectively. This allows the institution to build its knowledge and share best<br />

practices (Sallis and Jones 2002, 78). CoPs can also be established by management (Thorpe 2003, 9)<br />

74


Sheryl Buckley and Apostolos Giannakopoulos<br />

rather than form naturally. While management cannot command CoPs to form, they can facilitate their<br />

growth (Thorpe 2003, 9).<br />

The perpetuation of a learning culture depends on the institution’s ability to develop practices, procedures<br />

and customs that draw new people into a CoP and compensate for membership losses (Buysse,<br />

Sparkman and Wesley 2003, 267).<br />

White and Weathersby’s (2005, 294) view is that the traditional culture of educational institutions often<br />

runs counter to the values of the learning organisation. <strong>Academic</strong>s value academic freedom, intellectual<br />

development, exploration and examination, peer review, acquisition of knowledge for its intrinsic worth,<br />

and collegiality. <strong>Academic</strong>s are also members of a loosely-coupled community, working as experts in<br />

their self-defined domain of a larger field, collaborating when necessary, and governed increasingly by<br />

professional academic administrators.<br />

White and Weathersby (2005, 295) contend that educators have the opportunity of choosing to become<br />

members of multiple CoPs within their own organisation and larger profession. Creating learning-oriented<br />

CoPs among faculty in even small, incremental ways is a challenge that may involve risk taking.<br />

4. Research methodology<br />

The research on which this article is based aimed at investigating CoP awareness and views of<br />

academics at a higher education institution in SA. A mixed method approach was used. Qualitative data<br />

was aimed at gaining a better understanding of the awareness and views of academics using a CoP. The<br />

sample consisted of 784 academics. A questionnaire consisting of four parts (section A, B, C and D)<br />

including open- and closed-ended questions was administered. Section A contained background<br />

information and for the purpose of this article, this data will not be discussed. Section B consisted of five<br />

closed-ended questions and aimed to determine whether academics were prepared to share their<br />

knowledge and experience with others in the same field; whether the character of the other members<br />

would play a role; whether they preferred to work alone or were team players, and which form of<br />

communication they preferred. Section C consisted of 16 closed-ended questions informing participants<br />

about CoPs and establishing their willingness to take part in a CoP. Section D (open-ended) was an<br />

opportunity for participants to highlight any other factors not mentioned about CoPs in section B and C.<br />

The questionnaire was administered (by an independent agency) online with a weekly email reminder to<br />

the participants for the duration of three weeks. Only 180 participants completed the questionnaire and<br />

this could be a limitation. A reason could be that academics were busy with assessments and preparation<br />

of mid-year examinations during that period. The sample may therefore be considered too small to<br />

generalise but trends can be noted.<br />

Content analysis and open coding was used to analyse the qualitative data. The reliability of the<br />

quantitative section of the questionnaire was determined by using Cronbach’s Alpha test to determine<br />

whether the measuring instrument was homogeneous.<br />

5. Results and discussion<br />

The questionnaire was completed by 22.9% (180) of the respondents.<br />

Section B produced the following results: All respondents indicated that they would be prepared to share<br />

knowledge and experience with others in the same field. This is a positive sign when formulating a<br />

strategy for the implementation of CoPs. The character of the other members would play a role for 77.8%<br />

(140) while sharing knowledge. This could be indicative of trust or social issues experienced by some<br />

respondents. Muller (2006) argues that effective knowledge sharing requires investing in social<br />

relationships. While 88.3% (159) of the respondents were of the opinion that they were team players,<br />

43.3% (78) of the respondents preferred to work alone. This was surprising, as 88.3% (159) of the<br />

respondents had indicated previously that they felt they were team players. Face-to-face communication<br />

was preferred by 85.5% (154) of respondents and 72.7% (131) preferred online communication. This is<br />

yet another positive sign for implementing a CoP, because as stated earlier, a CoP is a group of people<br />

who come together to learn from one another, both face-to-face and virtually (Wenger 2004).<br />

Correspondence was also indicated as an option.<br />

Section C aimed at informing participants about a CoP and at determining whether they would be willing<br />

to participate (see Table 1).<br />

75


Table 1: Informing participants about a CoP<br />

*Legend for<br />

theme<br />

categories<br />

Sheryl Buckley and Apostolos Giannakopoulos<br />

Question Disagree Agree Strongly<br />

agree<br />

KS I benefit from knowledge sharing 3 107 70<br />

PD I will learn more from a colleague 57 93 24<br />

KS We all possess tacit knowledge 6 107 64<br />

KS I would like to discuss my teaching with others 90 57 9<br />

KS I would like to share my teaching methods 37 121 17<br />

BR Sharing my personal knowledge builds trust 11 108 61<br />

ATT Sharing my knowledge increases my status 21 108 50<br />

PD By sharing knowledge, knowledge base will increase 1 70 109<br />

MP Sharing knowledge will influence decisions 34 113 27<br />

ATT I have knowledge for my peers 16 135 28<br />

KS Sharing knowledge is not detrimental to performance 17 124 37<br />

PR CoPs are created out of passion and ‘die’ from lack of<br />

it<br />

23 116 35<br />

PI Same identity creates a strong bond 22 124 29<br />

VP Since voluntary, I can opt out anytime 26 116 33<br />

VP, BR Friendships can develop based on trust 6 118 56<br />

MP My employer benefits from knowledge sharing 1 110 65<br />

* See Table 2 for explanation of legend.<br />

Comments given by respondents in Section D were the following: ‘Employers that are senior should<br />

encourage knowledge sharing and reward us for sharing good ideas, they should not use it as their own<br />

platform of success’ and ‘I was not sure what you meant with CoP but having considered your definition<br />

and survey, I conclude that CoP principles were instrumental in advancing my research abilities and<br />

industry reputation.’ Other factors noted were excessive administrative requirements, fear of the<br />

unknown, time limitations, ignorance, lack of knowledge about CoPs, heavy workload, no promotion, nonexistence<br />

of a CoP, not being established in the institution, organisational culture and personal reasons.<br />

Section B, C and D were grouped and categorised into themes using content analysis and open coding<br />

(see Table 1 and Table 2) and are discussed below.<br />

Knowledge sharing: The desire to share knowledge was supported by the majority of respondents i.e.<br />

98.3% (177), who felt that they would benefit. Sharing knowledge was not seen as being detrimental to<br />

performance 91% (163) and 95% (171) agreed that we all possess tacit knowledge. This supports<br />

Wenger’s (2004) and Wick’s (2000, 520) findings. There were 76.6% (138) of the respondents who were<br />

willing to share their unique and effective teaching methods with others.<br />

Build relationships: Here again the majority of the respondents felt that they were team players. Muller<br />

(2006) contends that it takes time to invest in social relations. A small minority, namely 3.3% (6),<br />

disagreed that by associating voluntarily with others to share knowledge, friendships can develop, based<br />

on trust. Most respondents, namely 93.9% (169), agreed that sharing helps build trust among peers.<br />

Communication: Most respondents, namely 85.5% (154), also indicated that they would use face-to-face<br />

communication but 72.7% (131) indicated that they would prefer to communicate online. Communication<br />

was not a detrimental factor for the respondents.<br />

Personal development: The majority, namely 99.5% (179), agreed that by sharing knowledge, their<br />

knowledge base would increase. This view is supported by Leonard (1995, 63),Sallis and Jones (2002,<br />

78) and Wick (2000, 519). The results show that respondents are willing to socialise and in the end their<br />

knowledge base increases.<br />

Attitudes: Only a small percentage of the respondents, namely 11.7%, (21), disagreed that sharing<br />

personal knowledge and experiences would maintain or increase their status amongst their peers.<br />

Findings by Smith and McKeen (2004, 399) concur with these findings. There was a desire by 90.5%<br />

(163) to share knowledge with their peers. The results show that the respondents were willing to share.<br />

76


Sheryl Buckley and Apostolos Giannakopoulos<br />

Management participation: The majority of respondents, namely 97.2% (175), agreed that the<br />

organisation also benefits from the sharing of knowledge among the employees. This is supported by<br />

Sallis and Jones (2002, 36-37, 43) and Du Toit (2000, 188). Furthermore, 77.7% (140) agreed that<br />

sharing personal knowledge and experiences would increase their power to influence decisions. As is the<br />

case in any institution, management approval in an HEI needs to be granted before a CoP may be<br />

started especially if it will be institutionalised. Therefore management involvement and encouragement<br />

may be crucial.<br />

Voluntary participation: Almost full consensus was obtained among the respondents that by associating<br />

voluntarily with others to share knowledge, friendships can develop, based on trust. White and<br />

Weathersby (2005, 295) support this standpoint. However, 14.4% (26) felt that they could not opt out at<br />

any time. Perhaps they felt they would be obligated to stay in such a community once they were<br />

members. Since a CoP primarily implies voluntary participation, members may leave whenever they wish<br />

to do so.<br />

Practice: Most of the respondents, namely 83.8% (151), agreed that CoPs are created out of passion for<br />

one’s work. This is one of the characteristics of a successful CoP (Wenger 2004; Buysse et al. 2003, 267;<br />

Sherman et al. 1987, 72). The practice is the end goal of what has been derived from the learning<br />

between the interactions with others. We all possess tacit knowledge and through our interactions, our<br />

knowledge base is increased.<br />

Professional identity: Most of the respondents, namely 85% (153), agreed that sharing the same identity<br />

creates a strong bond among the members of a CoP. This is supported by previous research (Pee and<br />

Kankanhalli 2007, 134; Thorpe 2003; Barab and Duffy 1998). Wenger (2004) claims that the strong<br />

feeling of identity of members belonging to a CoP is mostly regarded as a positive aspect. Perhaps<br />

respondents in the same field feel comfortable to share with others in a similar field.<br />

Advertise: A respondent commented that participants could be informed about CoPs by advertising the<br />

existence of the CoP. It is important to promote CoPs continually to engage new members, re-engage<br />

existing members and sustain the community over its lifespan.<br />

These themes (see Table 2) were further grouped into three main themes (see Table 3): domain,<br />

community and practice, which form the elements of a CoP.<br />

Table 2: Initial categories of themes<br />

1 Management participation MP 8 Knowledge sharing KS<br />

2 Practice PR 9 Voluntary participation VP<br />

3 Communication COMM 10 Professional identity PI<br />

4 Building relationships BR 11 Advertise AD<br />

5 Attitudes ATT 12 Trust TR<br />

6 Personal development PD 13 Time TI<br />

7 Infrastructure INF<br />

Table 3: Final themes<br />

Domain 8<br />

Practice 1, 2, 6, 9, 10, 11, 13<br />

Community 3, 4, 5, 12<br />

77


Sheryl Buckley and Apostolos Giannakopoulos<br />

Factor analysis was carried out on the data of the survey to confirm the validity and reliability of the<br />

findings. The purpose of the factor analysis was to determine whether the 16 items could be organised or<br />

grouped into a smaller set of underlying factors. Four factors were generated from the 16 items and are<br />

presented in Table 4. For this article, 0.700 was used as the benchmark against which to measure the<br />

Cronbach’s Alpha values. The four factors that were generated were personal development (factor 1),<br />

mutual beneficiation (factor 2), self-esteem (factor 3) and self-recognition (factor 4).<br />

There was no noteworthy difference between the gender of the respondents for the four factors used. It is<br />

also possible that the respondents who participated in the survey met some of the criteria of a CoP.<br />

There was also no noteworthy difference between the ages, current positions or number of years<br />

employed of the respondents for the same factors used.<br />

Table 4: Factor analysis<br />

Factors generated Cronbach’s Alpha<br />

Personal development .740<br />

Mutual beneficiation .729<br />

Self-esteem .704<br />

Self-recognition .652<br />

It was concluded that the prerequisite conditions for CoP were present, so a framework for CoP was<br />

developed to help start and maintain CoPs. CoPs are formed by people who engage in a process of<br />

collective learning in a shared domain. The findings of this research confirm most of the findings of the<br />

literature that the three key elements that are critical for a CoP should be taken into consideration when<br />

creating a CoP, namely: domain, community and practice.<br />

Figure 1 presents a framework based on the themes specified in Table 2. These themes were grouped<br />

into the three key elements of a CoP: domain, community and practice. This framework could be used to<br />

assist in the creation of new CoPs and to improve existing CoPs where a society of CoPs can ultimately<br />

develop. The findings positively suggest that a CoP might increase knowledge sharing in a university.<br />

Domain has been linked to management, as management is perceived to have the resources to create<br />

awareness among the academic community, to educate, encourage use, provide support (financial and<br />

technical) and reward sharing. Without the support and approval of management, no CoP can be<br />

created. Community and practice have been linked to academics. <strong>Academic</strong>s could develop, use,<br />

support, mentor and evaluate CoPs. Finally, CoPs are integrated to form global societies of CoPs in<br />

which academics participate.<br />

6. Conclusion<br />

The main elements of a CoP either featured, or were accepted by most of the respondents. One factor<br />

that appears to inhibit either the flourishing of an existing CoP or the formation of a new one is too little<br />

time. Most of the respondents accepted the other elements. This is very encouraging, since if the<br />

necessary elements to establish a CoP are available, then half of the objective has already been<br />

achieved. The two factors that appear from the findings to play an important role in the effectiveness of<br />

knowledge sharing are the attitudes of the academics and participation by management. Management<br />

has the ability to create an atmosphere that can be conducive to the sharing and creating of knowledge<br />

among academics by focusing on the domain, community and practice.<br />

It is the combination of a domain, community and practice that constitute a community of practice. And it<br />

is by developing these three elements in parallel that such a community is cultivated. In a higher<br />

education institution, where knowledge is created, disseminated and transmitted, management perhaps<br />

needs to play an even more active role in developing communities of practice if such communities are to<br />

be successful.<br />

78


If existing<br />

CoP<br />

Management<br />

<strong>Academic</strong>s<br />

Create<br />

awareness<br />

(advertise<br />

existing and<br />

new CoPs)<br />

Use CoP<br />

Sheryl Buckley and Apostolos Giannakopoulos<br />

Support CoP<br />

Educate &<br />

encourage<br />

(workshops)<br />

Mentor CoP<br />

<strong>Academic</strong>s<br />

develop or<br />

build new<br />

CoPs<br />

Willingness to<br />

participate in a<br />

CoP<br />

Figure 1: Framework for the creation of a CoP (adapted from Buckley 2009, 178)<br />

Evaluate CoP<br />

Use CoPs<br />

Themes identified<br />

from questionnaire<br />

survey<br />

79<br />

Support CoP<br />

<strong>Academic</strong>s<br />

Integrate CoPs<br />

Knowledge sharing<br />

Mentor CoP<br />

Domain<br />

Attitudes<br />

Build relationships<br />

Community<br />

Evaluate CoP<br />

Create global and<br />

societies of CoPs<br />

Practice<br />

Integrate CoPs<br />

Advertise<br />

Communication<br />

Management participation<br />

Personal development<br />

Practice<br />

Professional identity<br />

Voluntary participation<br />

Time


References<br />

Sheryl Buckley and Apostolos Giannakopoulos<br />

Argyris, C. (1994). Good communication that blocks learning. Harvard Business Review 72(4):77-85.<br />

Barab, S.A. and Duffy, T.M. 1998. From practice fields to communities of practice.<br />

http://crlt.indiana.edu/publications/complete.pdf. (Accessed 7 December 2009).<br />

Brzezinski, L. (1997). Reflections on the university. In J. Brezezinski and L. Nowak (eds). The idea of the university.<br />

Amsterdam: Rodopi. Pp.201-221.<br />

Buckley, S.B. (2009). Knowledge sharing through communities of practice at institutions of higher education.<br />

University of Johannesburg.<br />

Buysse, V., Sparkman, K.L. and Wesley, P.W. (2003). Communities of practice: Connecting what we know with what<br />

we do. Exceptional Children 69(3):263-277.<br />

Chetley, A. and Vincent, R. (2003). Learning to share learning: An exploration of methods to improve and share<br />

learning. www.healthcomms.org. (Accessed 14 April 2010).<br />

Davidson, C. and Voss, P. (2002). Knowledge management: An introduction to creating competitive advantage from<br />

intellectual capital. Auckland: Tandem Press.<br />

Du Toit, A.S.A. (2000). Knowledge management: An indispensable component of the strategic plan of South African<br />

universities. South African Journal of Education 20(3):187-191.<br />

Galunic, C. and Weeks, J. (1999). Managing knowledge at Booz-Allen and Hamilton: Knowledge on-line and off.<br />

Fontainebleau, France: Instead.<br />

Gorelick, C., Milton, N. & April, K. (2004). Performance through learning: knowledge management in practice.<br />

Elsevier Butterworth-Heinemann.<br />

Hayes, N. and Walsham, G. (2000). Competing interpretations of computer supported co-operative work.<br />

Organization 7(1):49-67.<br />

Hislop, D. (2004). The paradox of communities of practice: Knowledge sharing between communities. In P.M.<br />

Hildreth & C. Kimble (Eds). Knowledge networks: Innovation through communities of practice. Hershey: Idea<br />

Group.<br />

Jacob, M. (2000). Imaging the future university. In M. Jacob and T. Hellström (Eds). The future of knowledge<br />

production in the academy. Buckingham: The Society for Research into Higher Education, Open University<br />

Press.<br />

Jarvis, P. (2007). Globalisation, lifelong learning and the learning society: Sociological perspectives. Lifelong learning<br />

and the learning society, Volume 2. Routledge.<br />

Lamberton, D.M. (2005). Information sharing. In D. Rooney, G. Hearn and A. Ninan. Handbook of the knowledge<br />

economy. Edward Elgar Publishing Limited.<br />

Lea, M.R. (2005). Communities of practice in higher education. Useful heuristic or educational model? In D. Barton<br />

and K. Tusting (Eds). Beyond communities of practice. Language, power and social context. Cambridge:<br />

Cambridge University Press.<br />

Leonard, D. (1995). Wellspring of Knowledge: Developing and Sustaining the source of Innovation. Boston: Harvard<br />

Business School Press.<br />

Liebowitz, J. and Chen, Y. (2003). Knowledge sharing proficiencies: The key to knowledge management. In C. W.<br />

Holsapple (Ed.), Handbook on knowledge management (pp. 408-438). New York: Springer-Verlag.<br />

Msila, F. (2006). Massification: Preparing faculty for open learning systems. Progressio 28(1, 2):82-96.<br />

Muller, P. (2006). Reputation, trust and the dynamics of leadership in communities of practice. Journal of<br />

Management and Governance 10(4):381-400.<br />

Pee, L.G. and Kankanhalli, A. (2007). Development of knowledge management capability in an institution of higher<br />

learning in Singapore. In K.S. Ning (Ed.). Knowledge management: Singapore perspectives 2007. Singapore:<br />

Information and Knowledge Management Society.<br />

Sallis, E. and Jones, G. (2002). Knowledge Management in Education. Enhancing Learning and Education. London:<br />

Kogan page Limited.<br />

Sherman, T.M., Armistead, L.P., Fowler, F., Barksdale, M.A. and Reif, G. (1987). The quest for excellence in<br />

university teaching. The Journal of Higher Education 58(1):66-84.<br />

Skyrme, D. (2003). Knowledge management: Making sense of an oxymoron. www.skyrme.com/insights/22km.htm.<br />

(Accessed 1 September 2009).<br />

Smith, H.A. and McKeen, J.D. (2004). Knowledge management in organisations: The state of current practice. In<br />

C.W. Holsapple (Ed.). Handbook on knowledge management 2. Knowledge directions (p. 395-410). Berlin:<br />

Springer-Verlag.<br />

Thorpe, M. (2003). Communities of practice and other frameworks for conceptualising, developing and evaluating<br />

NCSL’s initiatives in linking staff and school communities. Unpublished report for the National College of School<br />

Leadership.<br />

Wenger, E. (2004). Knowledge management as a doughnut: Shaping your knowledge strategy through communities<br />

of practice. Ivey Business Journal, January/February.<br />

http://www.iveybusinessjournal.com/view_article.asp?intArticle_ID=465. (Accessed 7 December 2009).<br />

Wenger, E. (2006). Communities of practice: A brief introduction.<br />

http://www.ewenger.com/theory/communities_of_practice_intro.htm. (Accessed 7 December 2009).<br />

White, J. and Weathersby, R. (2005). Can universities become true learning organizations? The Learning<br />

Organization 12(3):292-298.<br />

Wick, C. (2000). Knowledge management and leadership opportunities for technical communicators. Technical<br />

Communication 47(4):515-529.<br />

80


Sheryl Buckley and Apostolos Giannakopoulos<br />

Witt, N., McDermott, A., Peters, M. and Stone, M. (2007). A knowledge management approach to developing<br />

communities of practice amongst university and college staff.<br />

http://www.ascilite.org.au/conferences/singapore07/procs/witt.pdf. (Accessed 30 September 2009).<br />

81


An Empirical Framework of key Success Factors for<br />

Software Process Improvement<br />

Nuntarat Bunditwongrat, Mathupayas Thongmak and Atcharawan Ngarmyarn<br />

Thammasat University, Bangkok, Thailand<br />

nuntarat.b@gmail.com<br />

mathupayas@gmail.com<br />

a_ngarmyarn@yahoo.com<br />

Abstract: During the last decade, software industry has concerned more about the quality of software process, as<br />

indicated by an increasing number of proposed software process improvement (SPI) standards and models.<br />

However, the adoption rate of these standards and models in software development is still low and the failure rate of<br />

SPI implementations is very high. For these reasons, this paper is aimed at investigating the key success factors for<br />

SPI implementation to answer two research questions. The first question is, what factors do respondents think that<br />

they have an impact on SPI implementation success?; and the second question is, does the organizational size have<br />

moderating effect on the relationship between human factors / organizational factors and SPI implementation<br />

success? The conceptual model and hypotheses of this study are proposed by extending and integrating conceptual<br />

models from prior research, collecting the human, organizational, and implementation factors associated with SPI<br />

implementation in industry. Furthermore, the contextual factor that is organizational size was added to the model to<br />

test the sensitivity of the variation in software organizations. Then, a quantitative questionnaire survey was gathered<br />

from 23 software organizations in Thailand which apply CMMI (Capability Maturity Model Integration) as their SPI<br />

guidelines in order to test the conceptual model and hypotheses of the study. The results indicate that SPI<br />

implementation success depends on six factors, i.e. management commitment, staff involvement, allocated<br />

resources, pilot projects, setting SPI goals, and defined process model, while organizational size has no moderating<br />

effect on SPI implementation success. Finally, the main contribution of this paper is the guidance for SPI practitioners<br />

on key factors that practitioners need to focus during an SPI implementation. Moreover, this guidance will be useful<br />

in the processes of allocating resources and prioritizing tasks to effectively implement SPI in the organization.<br />

Keywords: software process improvement, SPI, CMMI, key success factors<br />

1. Introduction<br />

Software plays an increasing role in today’s society. Modern software organizations operate in a highly<br />

dynamic market, under time and cost constraints (Abrahamsson 2001). Software quality has received<br />

attention from both academic and software industry (Niazi, Wilson, and Zowghi 2003a, 2003b).<br />

Organizations have started to apply SPI initiatives to increase the maturity and quality of their software<br />

processes (Abrahamsson 2001).<br />

During the last decade, the software industry has concerned more about software process improvement<br />

as indicated by an increasing number of proposed standards and models. All of these standards are<br />

claimed to increase the likelihood of SPI success (Dyba ํ 2001, 2005). Investment in process improvement<br />

has significant benefits such as improving productivity and product quality, reducing cost and time to<br />

market, increasing organizational flexibility and customer satisfaction, and enabling opportunities to enter<br />

the international market (Abrahamsson 2001; Brietzke and Rabelo 2006; Jalote 2002; Niazi, Wilson, and<br />

Zowghi 2003b).<br />

The number of proposed SPI standards and models has been increased; however, there has been<br />

limited success for many SPI efforts. Several companies have been carrying out SPI projects but some of<br />

them give up before realizing its benefits and others take much longer than expected to get it<br />

accomplished. The current problem with SPI is a lack of an effective strategy to successfully implement<br />

those standards or models (Abrahamsson and Iivari 2002; Babar and Niazi 2008; Brietzke and Rabelo<br />

2006; Hardgrave and Armstrong 2005; Niazi, Wilson, and Zowghi 2003b, 2005).<br />

CMMI is currently the most widely used framework for SPI. The recent report from SEI indicated that<br />

there was an increasing in CMMI appraisal from software companies around the world up to 15% during<br />

the last six months of 2008 (ThailandSPIN 2009). Thailand is the second highest number of appraised<br />

companies in Southeast Asia; however, the number is still low compared with all software companies in<br />

Thailand (Poopaka 2009; Siamturakij Media 2009). So, it is interesting to investigate the key success<br />

factors for SPI to assist software companies in effectively planning SPI implementation strategies. In this<br />

paper we focus on two research questions:<br />

82


Nuntarat Bunditwongrat et al.<br />

RQ1: What factors do respondents think that they have an impact on SPI implementation success?<br />

RQ2: Does the organizational size have moderating effect on the relationship between human<br />

factors/organizational factors and SPI implementation success?<br />

2. Background<br />

2.1 SPI concept<br />

During the last decade, SPI has become a popular approach to deliver improvements in software<br />

products. Many companies have either a formal or informal SPI program based on SPI model, such as<br />

CMM, CMMI, and ISO/IEC 15504 (SPICE). These models are SPI frameworks for defining and<br />

measuring processes and practices that can be used by software companies (Babar and Niazi, 2008;<br />

Hall, Rainer, and Baddoo 2002; Niazi, Wilson, and Zowghi 2003b). The report from Data & Analysis<br />

Center for Software (DACS) support that successful SPI programs reduce number of defects by 95%,<br />

reduce software development schedules by 71%, and increase productivity in terms of lines-of-code or<br />

function points per day by 222%. Additionally, Software Engineering Institute (SEI) reported the ratio of<br />

average return of 5:1 for investments in successful SPI programs (Abrahamsson 2001).<br />

2.2 CMMI<br />

CMMI is a model used to develop and refine software development processes which measures both<br />

capabilities and maturity levels. The model was developed by SEI, University of Carnegie Mellon, USA.<br />

CMMI is the successor of the CMM which integrate and standardize the separate models of CMM, and to<br />

eliminate other drawbacks of CMM. It is accepted and widely used in software development over the<br />

world (Malaiwong 2004, 2008a).<br />

CMMI identified processes to efficiently improve software quality starting from design, delivery, and<br />

maintenance. The CMMI version 1.2 have five maturity levels consisting of 25 process areas which is the<br />

group of best practices needed to comply to achieve those task objectives. Higher maturity levels require<br />

more set of process areas to be complemented (Phillips 2005; National Electronics and Computer<br />

Technology Center 2009).<br />

The result from survey indicates that software organizations which apply CMMI as their SPI guideline can<br />

work more systematic and easier to track with the flexible standard practices. By using these standard<br />

practices, managers need to seek an adaptive strategy to deal with any changes that may occur all the<br />

times. As a result, managers can see working situation more clearly in order to solve problems efficiently<br />

and well prepare to solve unexpected problems that may occur (Malaiwong 2008b).<br />

3. Conceptual model and hypotheses<br />

The conceptual model and hypotheses of the study are proposed by extending and integrating<br />

conceptual models from prior research. We focus on the human, organizational, and implementation<br />

aspects. Also, the contextual factor was added to the model to test the sensitivity of the variation in<br />

organizational size. Figure 1 shows the research model.<br />

3.1 Human factors<br />

3.1.1 Management commitment<br />

Commitment concerns with the extent to which resources are made available for SPI and management's<br />

interest in SPI (EI-Emam, Goldenson, McCurley, and Herbsleb 2001). Without commitment from all<br />

organizational levels, the initiative will likely fail or the results are not reached. The concept of<br />

commitment has been brought up as one of the most important human factors determining whether a<br />

well-planned process improvement program will succeed or not (Abrahamsson 2001).<br />

SPI initiatives involve a large number of people. Overall, SPI initiative does not get support unless<br />

management shows strong commitment to provide both people and financial resources for the initiative,<br />

and to spend their own time in participating, monitoring, and resolving issues. Active involvement of<br />

management provides strong motivation to the people participating in the initiative (Jalote 2002). In<br />

addition, Steltzer and Mellis (1999) ranked management commitment as number one of the ten factors<br />

83


Nuntarat Bunditwongrat et al.<br />

for SPI success. Management should also show that they believe in the initiative. Thus, we hypothesize<br />

that<br />

H1: SPI success is positively associated with management commitment<br />

3.1.2 Staff involvement<br />

Staff involvement is essential because if staff do not buy into the proposed changes, the improvement<br />

initiative is useless. Furthermore, they have detailed knowledge of the current processes. Steltzer and<br />

Mellis (1999) also ranked staff involvement as number two of important success factors with two reasons.<br />

First, their first hand experience of the processes is valuable. Second, participation from staff reduces the<br />

resistance to the developed processes.<br />

While Dyba ํ (2003, 2005) defines staff involvement as the extent to which employees use their<br />

knowledge and experience to decide, act, and take responsibility for SPI. Staff participation should be<br />

offered and managed to allow them improving their work and feeling a sense of contribution. Thus, we<br />

hypothesize that<br />

H2: SPI success is positively associated with staff involvement<br />

Human Factors:<br />

Management Commitment<br />

Staff Involvement<br />

Organizational Factors:<br />

Allocated Resources<br />

Pilot Projects<br />

Implementation Factors:<br />

Setting SPI goals<br />

Defined Process Model<br />

Figure 1: Conceptual research model<br />

3.2 Organizational factors<br />

3.2.1 Allocated resources<br />

Moderating Factor:<br />

- Organizational Size<br />

H7A H7B H7C H7D<br />

H5<br />

H6<br />

H1<br />

H2<br />

H3<br />

H4<br />

SPI Success<br />

Allocated resources are defined as the amount of staff time and resources dedicated to SPI (Goldenson,<br />

and Herbsleb 1995). SPI requires planning, dedicated people, management time, tools, and capital<br />

84


Nuntarat Bunditwongrat et al.<br />

investment. Organizations must commit their resources and constantly motivate people to contribute<br />

(Borjesson and Mathiassen 2004). SPI can’t be sustained without dedicated resources. Shen and Ruan<br />

(2008) recommend that at least one full time person for the SPI project.<br />

Training may be considered as a tool for implementing a successful SPI program and it requires allocated<br />

resources as well (Niazi, Wilson, and Zowghi 2003a, 2005). Training involves transferring important<br />

information and communicating process innovations to all relevant members. Providing enhanced<br />

understanding is mentioned in 72% of the ISO cases and in 87% of the CMMI cases (Stelzer and Mellis<br />

1999). It is important for capturing and disseminating experiences to less experienced staff (Rainer and<br />

Hall 2002). Thus, we hypothesize that<br />

H3: SPI success is positively associated with allocated resources<br />

3.2.2 Pilot projects<br />

After the process model was approved, the newly revised processes went through a testing phase (Shen<br />

and Ruan 2008). A careful planning to pilot new processes is critical because only some projects will be<br />

selected for pilot implementation. Good pilot planning is essential to maintain on-time SPI initiatives<br />

(Serrano, Montes, and Cedillo 2003).<br />

Practitioners are advised to implement SPI program within a particular department first. After proper<br />

planning and using pilot implementation experience, results of pilot implementation should be illustrated<br />

to other departments to get support and confidence (Niazi, Wilson, and Zowghi 2003b, 2005). Thus, we<br />

hypothesize that<br />

H4: SPI success is positively associated with pilot projects<br />

3.3 Implementation factors<br />

3.3.1 Setting SPI goals<br />

Stelzer and Mellis (1999) defined SPI goals as the degree to which everyone can make the translation<br />

from top management goals to the goals that each person is being asked to achieve. The goals must be<br />

decomposed to specific measures that everyone can see how their efforts relate to organization's<br />

success. Setting relevant goals means that the improvement efforts attempt to contribute to the success.<br />

While setting realistic goals means that the goals may be achieved in the foreseeable future with a<br />

reasonable amount of resources. This helps to direct the efforts towards common goals and to motivate<br />

people.<br />

The role of process improvement has been recognized as essential factor to achieve business objectives.<br />

So, clearly defined SPI program driven by business needs, have been suggested as a key success factor<br />

for SPI (Dyba ํ 2005). Thus, we hypothesize that<br />

H5: SPI success is positively associated with setting SPI goals<br />

3.3.2 Defined process model<br />

The major activities of defining process model involve in some definition of new processes and refine of<br />

existing processes. Frequently, process definition is the main challenge in an SPI initiative. Besides, it is<br />

imperative to have deployment mechanisms in place for deploying the processes to make the task of<br />

changing processes easier (Jalote 2002; Niazi, Wilson, and Zowghi 2003a; Shen and Ruan 2008).<br />

Creating process action team is also considered as a mechanism to effectively change or deploy new<br />

processes in the organizations. It is recommended to have a core group of full-time people that may<br />

include external agents (Hall, Rainer, and Baddoo 2002; Jalote 2002). The rest of the team should be<br />

users in processes, e.g. project managers, developers, testers, etc, to minimize the risk of resistance<br />

(Shen and Ruan 2008). Thus, we hypothesize that<br />

H6: SPI success is positively associated with defined process model<br />

85


3.4 Organizational size<br />

Nuntarat Bunditwongrat et al.<br />

Existing software engineering and organization development literature acknowledges that there are<br />

fundamental operational differences between small and large organizations. Small organizations seem<br />

more concerned about practice, while large organizations seem more concerned about formal process<br />

(Dyba ํ 2003). Despite this recognition, there has been no systematic attempt to verify whether small and<br />

large software organizations implement SPI programs differently. So, it is interesting to verify the<br />

importance of organizational size on SPI success in this study. Thus, we hypothesize that<br />

H7A: Organizational size moderates the effect of management commitment on SPI success.<br />

Management commitment influences the SPI success for large organization to a greater extent than that<br />

of small organization.<br />

H7B: Organizational size moderates the effect of staff involvement on SPI success. Staff involvement<br />

influences the SPI success for large organization to a greater extent than that of small organization.<br />

H7C: Organizational size moderates the effect of allocated resources on SPI success. Allocated<br />

resources influence the SPI success for large organization to a greater extent than that of small<br />

organization.<br />

H7D: Organizational size moderates the effect of pilot projects on SPI success. Pilot projects influence<br />

the SPI success for large organization to a greater extent than that of small organization.<br />

3.5 SPI success<br />

Dyba ํ (2003, 2005) measured SPI success based on the performance of the organization for the past<br />

three years with respect to cost reduction, cycle time reduction, and customer satisfaction. While<br />

Goldenson, Gibson, & Ferguson (2004) identified seven categories of performance measures: process<br />

adherence, cost, schedule, productivity, quality, customer satisfaction, and return on investment.<br />

Goldenson and Herbsleb (1995) asked the survey respondents how they would describe their<br />

organizations with respect to six performance characteristics: ability to meet schedule and budget<br />

commitments, address process predictability, product quality, staff productivity, staff morale / job<br />

satisfaction, and customer satisfaction. The overall patterns are quite clear that higher process maturity<br />

does appear to pay off in better organizational performance.<br />

4. Research design<br />

The data was collected between March and April 2010 from target sample of 293 employees who involve<br />

in SPI program of 23 software companies listed by Software Park, an agency supporting software<br />

companies in Thailand, using purposive sampling method. The questionnaires were distributed through<br />

both online and offline channels.<br />

The appraised companies comprise with the group of companies in maturity level 2, level 3, and level 5.<br />

There is no company in Thailand which achieved at level 4. According to statistics during the last two<br />

years, the appraised companies in Thailand has been increased up to 4 times, total of 30 companies<br />

divided into 9, 20, and 1 companies which achieve CMMI maturity level 2, level 3 and level 5 respectively.<br />

Notwithstanding, after contacting to the company which is mentioned to level 5 according to the list, we<br />

found that this company has not achieved at level 5, they are still at level 3. Hence, the sample of this<br />

research will cover only those companies achieved at level 2 and 3.<br />

5. Data analysis and result<br />

Our measurement model includes 40 items describing six latent constructs. The reliability of the multipleitem<br />

measurement scales were evaluated using Cronbach’s alpha coefficient. All constructs have a<br />

higher composite reliability than the benchmark of 0.7 recommended by Nunally (1978) as shown in<br />

Table 1.<br />

The unidimensionality of the six measurement scales was evaluated by performing an exploratory factor<br />

analysis using principal components analysis with varimax rotation. The six components were extracted<br />

86


Nuntarat Bunditwongrat et al.<br />

with eigenvalues greater than 1. The rotate matrix component is shown in Table 2. All item loadings for<br />

each component were greater than 0.5, which achieved an acceptable degree of unidimensionality.<br />

Table 1: Reliability analysis<br />

Independent Variables Number of Items α<br />

Management Commitment 6 0.847<br />

Staff Involvement 5 0.866<br />

Allocated Resources 10 0.939<br />

Pilot Projects 5 0.867<br />

Setting SPI Goals 6 0.893<br />

Defined Process Model 8 0.933<br />

Table 2: Rotated component matrixa<br />

Measurement Items<br />

1<br />

Component<br />

* 2 * 3 * 4 * 5 * 6 *<br />

A process has been established to distribute lessons<br />

learned to all relevant staff<br />

.734<br />

SPI implementation plan has been established .729<br />

The processes for improvement are clearly defined .721<br />

Mechanism has been established to manage changes .719<br />

Mechanism has been established to collect and analyze<br />

feedback data and to extract lessons learned<br />

.716<br />

Processes are adequately defined at an appropriate level<br />

of detail<br />

.714<br />

Continuously improve a SPI implementation methodology .664<br />

Mechanism has been established to monitor the progress<br />

of each process action team<br />

.576<br />

Sufficient budget for staff to participate in SPI training .722<br />

SPI training is provided for all related staff .698<br />

Sufficient time for staff to participate in SPI training .681<br />

Required tools have been provided for SPI implementation<br />

properly<br />

.673<br />

Required people have been provided for SPI<br />

implementation properly<br />

.663<br />

Sufficient staff time is allocated to SPI efforts .649<br />

SPI training helps developing skills and knowledge needed<br />

to perform SPI implementation<br />

.608<br />

Sufficient budget has been allocated to SPI implementation<br />

properly<br />

.571<br />

Training programs are reviewed on a periodic basis .537<br />

All future group or individual trainings of SPI are planned .520<br />

Staff members actively involve in defining their operational<br />

processes<br />

.780<br />

Staff members actively involve in setting goals for SPI<br />

activities<br />

.780<br />

Staff members have responsibility to involve in SPI<br />

activities<br />

.683<br />

Staff members interest and commit in SPI effort .675<br />

Opportunity for staff to exchange their ideas related to SPI<br />

implementation<br />

.549<br />

Management provides strong commitment for SPI .760<br />

Management actively participates in SPI activities .742<br />

Management actively supports SPI activities .679<br />

Management commits to provide resources for SPI .675<br />

Management establishes standard practices for SPI .638<br />

Management actively monitors the progress of SPI .578<br />

The results of pilot implementation are used for planning<br />

SPI implementation across the organization<br />

.664<br />

Careful planning of the piloting new processes .649<br />

Staff members satisfy with the performance of<br />

methodology in the pilot projects<br />

.646<br />

SPI implementation methodology has been tried and tested<br />

in pilot projects<br />

.567<br />

87


Nuntarat Bunditwongrat et al.<br />

Measurement Items<br />

1<br />

Component<br />

* 2 * 3 * 4 * 5 * 6 *<br />

The results of pilot implementation are communicated .540<br />

SPI goals are measurable .660<br />

SPI goals are aligned with organization’s business goals .607<br />

SPI goals are realistic .589<br />

SPI goals are well understood within the organization .589<br />

Fine balance between short-term and long-term SPI goals .544<br />

SPI goals are clearly stated .524<br />

Cronbach’s Alpha .847 .866 .939 .867 .893 .933<br />

a. Rotation converged in 7 iterations.<br />

*1 - Defined Process Model *2 - Allocated Resources *3 - Staff Involvement<br />

*4 - Management Commitment *5 - Pilot Projects *6 - Setting SPI Goals<br />

For respondent demographics, the ratio of female (51.54%) is slightly more than male (48.46%). The<br />

majority of the respondents’ age ranged between 25-35 years. Most of them have been working in the<br />

jobs related to software development process for 1-2 years. Most of the respondents’ company have staff<br />

dedicated to SPI in the range of 50-200 people with CMMI maturity level 3 and has started SPI<br />

implementation for 1-2 years.<br />

Multiple regression analysis was used to test the hypotheses. The regression model was run without<br />

constant because we used factor score to run in the model. Hence, the unstandardized and standardized<br />

coefficients were equal. Hypotheses 1 through 6 consider the individual relationships between SPI<br />

success and each of the six independent variables to answer our first research question. Table 3<br />

summarizes the results of our analysis for these hypotheses. The results show that hypotheses 1 through<br />

6 were significant. So, all six independent variables investigated have an impact on SPI implementation<br />

success.<br />

The contextual factor was added to the model to test the sensitivity of the variation in organizational size<br />

which is in accordance to our second research question. However, the estimated coefficients of the<br />

independent variables of the original model were not significantly changed by the introduction of the<br />

organization size. Thus, our analysis rejects Hypotheses 7A through 7D.<br />

Table 3: Regression analysis for hypotheses H1 – H6<br />

Model Standardized Coefficients (Beta) t<br />

Management Commitment .141 2.233*<br />

Staff Involvement .265 .757***<br />

Allocated Resources .218 2.514*<br />

Pilot Projects .229 .715**<br />

Setting SPI Goals .177 1.977*<br />

Defined Process Model .197 2.281*<br />

Dependent Variable: SPI Success<br />

R = .514 R2 = .265 SE = 0.865 (F = 17.202***)<br />

*p < .05 **p < .01 ***p < 0.005<br />

6. Discussion<br />

The results show that SPI success depends on six factors. Staff involvement seems to be associated with<br />

the highest explanatory power since it achieved the highest standardized regression coefficient ( =<br />

0.265). Next come to pilot projects ( = 0.229), followed by allocated resources ( = 0.218), defined process<br />

model ( = 0.197), setting SPI goals ( = 0.177), and finally, management commitment ( = 0.141). A<br />

88


Nuntarat Bunditwongrat et al.<br />

surprising result was the lowest explanatory power of management commitment in predicting SPI<br />

success. All the main studies on which this investigation is based share a strong belief in the importance<br />

of management commitment for the successful implementation of any type of change initiatives.<br />

However, the result is in agreement with Abrahamsson (2000)’s findings that many SPI initiatives do not<br />

require management commitment beyond obtaining the resources needed. It can be stated that if a<br />

committed manager should bring energy, passion, and excitement to improvement initiative, then<br />

management commitment is not a necessity. On the other hand, if management commitment is viewed<br />

as someone who provides resources that enables the process improvement to grow, then the<br />

management commitment is a necessity. Furthermore, the success in some cases as presented in<br />

Abrahamsson’s study could be the concept of champion, who makes a decisive contribution to the<br />

innovation by actively and enthusiastically making the improvement happen with having little support from<br />

management. There could be several explanations for this seemingly surprising result.<br />

With respect to the contextual factor, we found that organizational size did not have moderating effect on<br />

the relationship between human factors / organizational factors and SPI implementation success.<br />

However, the result is in agreement with Dyba ํ (2005)’s findings that the results hold for small, as well as<br />

large, software organizations, and for organizations operating in stable as well as in turbulent<br />

environments. Similarly, Goldenson and Herbsleb (1995) found that organizational size alone is unrelated<br />

to overall SPI success. Respondents who represent larger organizations are no more, or less, likely to<br />

claim such success than are those who come from smaller organizations. This means that the size of the<br />

organization does not limit its potential for SPI success. We believe that the success factors discussed<br />

here give many interesting insights that will be relevant and applicable to organizational change in most<br />

software organizations. If the success factors are implemented, they facilitate the success of<br />

improvement initiatives. If they are not implemented or not implemented correctly, this makes process<br />

improvement difficult to achieve, or may even cause failure of the initiative.<br />

7. Limitations<br />

Our participants belonged to only one standard, CMMI, and only one country, Thailand, which is a<br />

limitation of this study as the findings cannot be widely generalized to practitioners using other standards<br />

or models, and for other countries that have different environment and culture. As mentioned earlier, the<br />

sample of this research covers only those companies achieved at level 2 and 3. Nevertheless, software<br />

companies in Thailand are very productive and it would be another interesting case to investigate since<br />

they are well supported by many government agencies as witnessed in the statistics that the appraised<br />

companies in Thailand have been increased up to 4 times during the last two years (Malaiwong 2004).<br />

Another issue is that the questionnaire-based studies are usually based on self-reported data that may<br />

reflect what people think they should say, not necessarily what they actually observe or perceive. Our<br />

results are limited to the respondents’ knowledge and beliefs about the reasons, potential benefits and<br />

barriers that can undermine SPI initiatives. This situation can cause problems when practitioners’<br />

perceptions may be inaccurate. However, like many other studies based on opinion data, we have<br />

collected data from practitioners working in quite diverse roles and directly involved in SPI activities within<br />

their organizations to ensure that they could be representatives for our target population.<br />

8. Conclusion and future research<br />

The study focused on identifying the key factors for success in SPI by a quantitative survey. The results<br />

indicate that SPI implementation success depends on six factors, i.e. management commitment, staff<br />

involvement, allocated resources, pilot projects, setting SPI goals, and defined process model. While<br />

organizational size has no moderating effect on SPI implementation success. From a theoretical<br />

perspective, these findings add an important new dimension to empirical software engineering research<br />

to verify the importance of organizational size for SPI success. From a practical perspective, the main<br />

contribution of this paper is an actionable guidance for SPI practitioners on key factors that practitioners<br />

need to manage during an SPI effort. This guidance will be useful in processes of planning and<br />

prioritizing the implications, so that software organizations can evaluate their readiness and current<br />

practices to effectively implement SPI. Future research is needed to investigate such variable in other<br />

countries against the results of this study to see whether the different cultures and/or environments have<br />

any effect to the results. Furthermore, in case that the sample can be separated into more variety group<br />

of maturity levels, it is interesting to investigate the relationship between the success factors and maturity<br />

levels. Consequently, we would know that do the respondents from lower and higher maturity companies<br />

think that the same factors impact the implementation of SPI. Likewise, we can compare the factors<br />

recognized by companies with successful and non-successful SPI programs, as well as comparing CMMI<br />

to other standards/models, e.g. ISO/IEC 15504 (SPICE) or ITIL. In addition, qualitative research by in-<br />

89


Nuntarat Bunditwongrat et al.<br />

depth interview or case study should be done to gather insight into the SPI implementation success on<br />

how to efficiently establish SPI implementation strategies.<br />

References<br />

Abrahamsson, P. (2000) “Is Management Commitment a Necessity After All in Software Process Improvement?”,<br />

Proceedings of the 2<strong>6th</strong> Euromicro <strong>Conference</strong>, Vol. 2, pp 246-253.<br />

Abrahamsson, P. (2001) “Commitment Development in Software Process Improvement: Critical Misconceptions”,<br />

23rd International <strong>Conference</strong> on Software Engineering, pp 71-80.<br />

Abrahamsson, P. and Iivari, N. (2002) “Commitment in Software Process Improvement In Search of the Process”,<br />

Proceedings of the 35th Annual Hawaii International <strong>Conference</strong> on System Sciences, Vol. 8, pp 1-10.<br />

Babar, M.A. and Niazi, M. (2008) “Implementing Software Process Improvement Initiatives: An Analysis of<br />

Vietnamese Practitioner’s Views”, 2008 IEEE International <strong>Conference</strong> on Global Software Engineering, pp 67-<br />

76.<br />

Borjesson, A. and Mathiassen, L. (2004) “Successful Process Implementation”, IEEE Software, Vol. 21, No. 4, pp 36-<br />

44.<br />

Brietzke, J. and Rabelo, A. (2006) “Resistance Factors in Software Process Improvement”, CLEI ELECTRONIC<br />

JOURNAL, Vol. 9, No. 1.<br />

Dyba ํ, T. (2001) “An Instrument for Measuring the Key Factors of Success in Software Process Improvement”,<br />

Empirical Software Engineering, Vol. 5, No. 4, pp 357-390.<br />

Dyba ํ, T. (2003) “Factors of Software Process Improvement Success in Small and Large Organizations: An Empirical<br />

Study in the Scandinavian Context”, Foundations of Software Engineering, pp 148-157.<br />

Dyba ํ, T. (2005) “An Empirical Investigation of the Key Factors for Success in Software Process Improvement”, IEEE<br />

TRANSACTIONS ON SOFTWARE ENGINEERING, Vol. 31, No. 5, pp 410-424.<br />

EI-Emam, K., Goldenson, D., McCurley, J. and Herbsleb, J. (2001) “Modelling the Likelihood of Software Process<br />

Improvement: An Exploratory Study”, Empirical Software Engineering, Vol. 6, No. 3, pp 207-229.<br />

Goldenson, D.R. and Herbsleb, J.D. (1995) “After the Appraisal: A Systematic Survey of Process Improvement, its<br />

Benefits, and Factors that Influence Success”, Software Engineering Institute, pp 1-50.<br />

Goldenson, D.R., Gibson, D.L. and Ferguson R.W. (2004) “Why Make the Switch? Evidence about the benefits of<br />

CMMI”, Software Engineering Institute, pp 1-42.<br />

Hall, T., Rainer, A. and Baddoo, N. (2002) “Implementing Software Process Improvement: An Empirical Study”,<br />

Software Process Improvement and Practice, Vol. 7, pp 3-15.<br />

Hardgrave, B.C. and Armstrong, D.J. (2005) “Software Process Improvement: It’s a Journey, Not a Destination”,<br />

Communications of the ACM, Vol. 48, No. 11, pp 93-96.<br />

Jalote, P. (2002) “Lessons Learned in Framework-Based Software Process Improvement”, Proceedings of Ninth<br />

Asia-Pacific Software Engineering <strong>Conference</strong>, pp 1-5.<br />

Malaiwong, K. (2004) “Transition to CMMI”, [online], www.drkanchit.com/cmm/cmmitransition.pdf.<br />

Malaiwong, K. (2008a) “CMMI Implementation”, [online], www.drkanchit.com/cmm/CMMI_ ImplementationNew.pdf.<br />

Malaiwong, K. (2008b) “Benefits of CMM”, [online], www.drkanchit.com/cmm/cmm04.html.<br />

National Electronics and Computer Technology Center (2009) “Introduction to CMMI”, [online],<br />

cmmi.wikidot.com/knowledge.<br />

Niazi, M., Wilson, D. and Zowghi, D. (2003a) “A maturity model for the implementation of software process<br />

improvement: an empirical study”, The Journal of Systems and Software, Vol. 74, No. 2, pp 155-172.<br />

Niazi, M., Wilson, D. and Zowghi, D. (2003b) “A model for the implementation of software process improvement: a<br />

pilot study”, Proceedings of the Third International <strong>Conference</strong> on Quality Software, pp 196-203.<br />

Niazi, M., Wilson, D. and Zowghi, D. (2005) “A framework for assisting the design of effective software process<br />

improvement implementation strategies”, The Journal of Systems and Software, Vol. 78, No. 2, pp 204-222.<br />

Nunnally, J. C. (1978) Psychometric theory (2nd ed.), McGraw-Hill, New York.<br />

Phillips, M. (2005) “CMMI® Version 1.2 and Beyond …a Tutorial”. [online], Carnegie Mellon University,<br />

www.dtic.mil/ndia/2005cmmi/monday/phillips.pdf.<br />

Poopaka, S. (2009) “SPI@ease: CMMI Achievement Series #6 “CMMI vs Agile: Why Not Embrace Both !!!”,<br />

[seminar], Software Park Thailand.<br />

Rainer, A. and Hall, T. (2002) “Key success factors for implementing software process improvement: a maturitybased<br />

analysis”, The Journal of Systems and Software, Vol. 62, pp 71-84.<br />

Serrano, M.A., Montes, C. and Cedillo, K. (2003) “An Experience on using the Team Software Process for<br />

Implementing the Capability Maturity Model for Software in a Small Organization”, Proceedings of the Third<br />

International <strong>Conference</strong> On Quality Software, pp 327-334.<br />

Shen, B. and Ruan, T. (2008) “A Case Study of Software Process Improvement in a Chinese Small Company”, 2008<br />

International <strong>Conference</strong> on Computer Science and Software Engineering, Vol. 2, pp 609-612.<br />

Siamturakij Media (2009) “TMC supports software development of Thai private organizations”, [online],<br />

www.siamturakij.com/home/news/display_news.php?news_id=%20413339 303.<br />

Stelzer, D. and Mellis, W. (1999) “Success Factors of Organizational Change in Software Process Improvement”,<br />

Software Process: Improvement and Practice, Vol. 4, No. 4, pp 227-250.<br />

ThailandSPIN. (2009) “CMMI-appraised software companies in Thailand has been increased 3 times”, [online],<br />

Software Park Thailand, www.thailandspin.com/%20News/tabid/71/articleType/ ArticleView/articleId/42/-3-<br />

.aspx.<br />

90


Assessing Information Management Competencies in<br />

Organisations<br />

Andy Bytheway<br />

Cape Peninsula University of Technology, Cape Town, South Africa<br />

bythewaya@cput.ac.za<br />

Abstract: The history of the management of information systems includes many ideas that were intended to simplify<br />

the complexities of the management task, but there is still a great deal of wasted investment that produces no<br />

significant benefits. Much of the thinking has been rational and structured, but it can be argued that structured<br />

thinking will not solve the problems presented by the ever-increasing scope and depth of information systems, the<br />

need for improved responsiveness and agility, and the need to deal with a range of requirements that are sometimes<br />

behavioural and sometimes legislative. Three of the more frequently cited frameworks for information management<br />

(Zachman, Henderson & Venkatraman, Ward), are briefly reviewed and found to have common characteristics. They<br />

are combined into a new, simple arrangement of the central (and critically important) ideas. This new framework has<br />

been used as the basis of a survey instrument that is introduced and explained; it works at two levels - the "micro"<br />

and "macro" levels. It assesses perceptions of organisational capability to manage information well, as seen by<br />

respondents who are normally employees working in different roles with varying responsibilities. The survey<br />

instrument comes with an analysis and reporting package that is found to be suitable for the needs of busy<br />

managers, and the way in which micro and macro data is presently analysed and presented is demonstrated using<br />

data from a reference dataset, a CIO workshop, an investigation within a real estate agency and a large financial<br />

services organisation. The contribution of this work to the research programme from which it emanated is<br />

summarised and future directions briefly explained.<br />

Keywords: Information management; perceptions; IS/IT strategy; alignment; assessment<br />

1. Background<br />

There is a history of difficulty in delivering benefits from information technology investments, and for the<br />

last 40 years experts have worked to ease the problems.<br />

During the 1990s business managers and academics strove to find answers to critical questions. An<br />

early, extensive, review of literature concerning information systems "success" (DeLone & McLean, 1992)<br />

was well received and has since been updated; specific research has looked at process-based<br />

collaboration across corporate boundaries (Bytheway & Braganza, 1997); academic attention focused on<br />

strategic alignment (Kearns & Lederer, 2000; Chan & Reich, 2007); concerns about "agility", ethics and<br />

alignment have emerged (Tallon et al., 2000; Sambamurthy et al., 2003; Symons, 2005).<br />

More recently, there has been increasing attention to the management of benefits (Chatterji, 2007; Ward<br />

& Daniel, 2005) but business newspapers still report problems (Anon, 2008) and there is still difficulty<br />

with enterprise-wide systems (Seddon et al., 2010). It is reported that information systems strategy is still<br />

not properly understood (Chen et al., 2010), and ethics (Mingers & Walsham, 2010) and behavioural<br />

issues (Beaudry & Pinsonneault, 2010) are of concern. There are staggering losses involved in<br />

information technology investments in the public sector (Anon, 2010).<br />

In the face of all this, how can the complexities of managing information technology and systems be dealt<br />

with? When costs of technology are spiralling, how can they be justified to senior business managers in<br />

the board room? Concepts of information technology strategy might mirror business strategy, but exactly<br />

what do the "alignment" of these IT and business strategies really mean?<br />

This paper briefly reviews some ideas from this history and derives a new framework that proves to be<br />

comprehensible and workable. The framework leads to a survey instrument that assesses organisational<br />

capability to manage information well.<br />

2. Some important ideas<br />

2.1 Zachman reveals complexity<br />

The complexities of information technology management were first revealed in the 1980s in a framework<br />

for information systems architecture (Zachman, 1987); Zachman took a broad view of the issues but his<br />

ideas were necessarily detailed, and his frequently cited six-by-six matrix, with layers of technology<br />

management down one side and different perspectives of the business across the top, was beyond many<br />

91


Andy Bytheway<br />

managers' willingness to work with. 36 different points of concern (the intersection of the six rows and six<br />

columns of the matrix) were just too many to handle. Although there has been subsequent reference to<br />

the Zachman framework (Frankel et al., 2003) it has failed to gain currency in the general context.<br />

2.2 Henderson and Venkatraman seek simplicity<br />

Others have sought simpler views of the problem, and one frequently cited example is the Henderson<br />

and Venkatraman framework. It relates business and IT issues at the internal and external levels using a<br />

two-by-two matrix. It shows that there is a need for "functional integration" between the business and IT<br />

domains, and that there is a need for "strategic fit" between the internal and external worlds. Each of the<br />

four quadrants embodied in the framework are detailed in turn: scope, competencies and governance in<br />

the external portion and processes, skills and infrastructure or architecture in the internal portion.<br />

Internal External<br />

Strategic fit<br />

Distinctive<br />

competencies<br />

Business strategy IT strategy<br />

Business<br />

scope<br />

Administrative<br />

infrastructure<br />

Business<br />

Governance<br />

Processes<br />

Skills<br />

Organisational infrastructure and processes<br />

Automation Linkage<br />

Functional integration<br />

Systemic<br />

competencies<br />

Technology<br />

Scope<br />

Achitectures<br />

IT Governance<br />

Processes<br />

Skills<br />

IS Infrastructure and processes<br />

Business Information technology<br />

Figure 1: The Henderson and Venkatraman framework, linking business and IT from the internal and<br />

external viewpoints (Henderson & Venkatraman, 1993)<br />

Questions arise from a close examination of this model. It is interesting to see that competencies and<br />

skills are included, but why are they in different parts of the framework? A skill can be seen as a low level<br />

thing ("I can work this computer") but a competency is something else ("I can use this computer to<br />

produce a useful econometric model"). The implication of "processes" being in both the internal business<br />

and internal IT quadrants is that business processes and IT processes must be functionally integrated,<br />

but how is that possible? A single IT process might contribute to a wide range of business processes;<br />

conversely a typical business process might depend on many information systems. Why do we have<br />

"administrative infrastructure" on the left, and "architectures" on the right?<br />

The Henderson and Venkatraman model is simple at first sight, but it leads to a range of questions and<br />

lacks the sort of elegance and symmetry that makes these things memorable. It implies dependencies<br />

and relationships between its conceptual components, but these are not immediately evident on a first<br />

reading.<br />

92


2.3 The issue of alignment<br />

Andy Bytheway<br />

Efforts to improve the return on information technology investments have focused on the ways in which<br />

business and technology strategies can be more effectively aligned. However, this high-level approach to<br />

the management of IT and IS related investments has found only limited success.<br />

It has been argued (Yayla & Hu, 2009) that alignment is achieved by means of simple steps: strategising<br />

processes, increasing the level of communication, formalising policies, and so on. Easy to say, but is this<br />

taking us forwards? Where is the detail that tells us how to realise these worthy and somewhat obvious<br />

steps to success?<br />

"Alignment" is an appealing word but it has to be judged by the benefits that are delivered. The<br />

management of benefits has been an issue from the very early days (Baets, 1992) right through to recent<br />

times; it is now dealt with in standard texts (Ward & Peppard, 2002; Ward & Daniel, 2005).<br />

2.4 The concept of value from IT investments<br />

"Alignment" is appealing, but "value" is more so, especially when talking to senior management.<br />

Venkatraman introduced the idea of "eras" (Venkatraman, 1994), his observations concerned the<br />

increasing scope and reach of information systems, the increasing degree of business change that was<br />

required to benefit from them, and the increasing value to be gained thereby.<br />

Inevitably, practice takes time to adopt and adapt to new ideas and practical concerns continued to be<br />

reported at about that time (Uchitelle, 1996). But what Venkatraman had given us was a clear message<br />

that we are concerned with the management of systems and the information that comprises the essence<br />

of those systems, not just the technology.<br />

Progressive organisations worked along these lines, and references to "IT management" were<br />

supplanted by references to "IS management", and then to "Information management", as in the case of<br />

BP Chemicals (Cross, 1995). There was academic attention to these ideas based on workshops with<br />

working managers (Bytheway, 1996), and one feature of the work at BP chemicals led ultimately to the<br />

framework that is presented here. John Cross dunned his idea "Jacob's ladder" - a management stairway<br />

to a strategic heaven?<br />

2.5 Jacobs' ladder<br />

The arrangement of the four steps in BP Chemical's "Jacob's ladder" is shown in the figure below: At the<br />

bottom is the technology that comprise infrastructure for systems and business activity, and at the top the<br />

business processes that serve the business strategy (Cross & Earl, 1997).<br />

Value<br />

creation<br />

Value<br />

realisation<br />

Business processes<br />

Information<br />

Applications<br />

Infrastructure<br />

In-house<br />

expertise<br />

Outsourced<br />

expertise<br />

Figure 2: Jacob's ladder as promoted in BP Chemicals<br />

It is interesting that the creation of value is seen at the top, with in-house expertise; the realisation of<br />

value is seen at the bottom, based on the use of outsourced expertise (outsourcing was one of the<br />

principal outcomes for BP Chemicals at this time).<br />

93


Andy Bytheway<br />

However, this notion of value generation is seen in other work. Zachman had already articulated six<br />

"levels", from the representation of technology infrastructure (program code and data definitions), through<br />

technology, systems, enterprise and context of the enterprise (Zachman, 1987); Venkatraman was<br />

promoting the idea that data, information, knowledge, action and result were all related in a similar<br />

progressive way (Venkatraman, 1996), and Ward has promoted the idea that benefits from IT<br />

investments are achieved at different levels (Ward & Daniel, 2005).<br />

These ideas are summarised in the table below, and related to the five levels of the Information<br />

Management Body of Knowledge (IMBOK) that are discussed in the next section.<br />

Table 1: Linking information technology to business strategy<br />

Zachman Venkatraman Ward IMBOK<br />

Highest Context Result Strategic<br />

objective<br />

Business strategy<br />

Enterprise model Action Functional benefit Business benefit<br />

Lowest Representation of<br />

system<br />

Knowledge Business change Business process<br />

System model Information Enabling change Information<br />

system<br />

Technology Data Technology Information<br />

technology<br />

Zachman's "representation of system" is really only of interest to technology management, but his other<br />

five levels show strong empathy with the other viewpoints. One can argue that there are actually five<br />

levels at which management must operate (from the lowest to the highest):<br />

Information technology must be acquired, configured, and used to provide the requisite infrastructure<br />

so that a business can store its data and operate its …<br />

Information systems, some of which will necessitate changes to the way that the business operates<br />

its …<br />

Business processes, wherein knowledge is deployed in order to initiate the actions that are expected<br />

to deliver the desired …<br />

Business benefits, that will in turn realise the organisation's …<br />

Business strategy.<br />

This arrangement of ideas strongly reflects the historical thinking that is available, and it has been<br />

adopted as the foundation of the Information Management Body of Knowledge, which currently<br />

comprises a knowledge base, a handbook, a developing survey instrument, and a supporting community<br />

web site (http://www.imbok.org).<br />

3. The Information Management Body of Knowledge<br />

As can be seen from the table above, at the heart of the IMBOK is the idea of the business process,<br />

incorporated by Zachman (with other ideas) in his enterprise model, seen by Venkatraman as the layer<br />

within which knowledge is deployed (in the taking of informed decisions), and by Ward as the level where<br />

business change is to be found. Zachmans system model is echoed strongly in the IMBOK, where<br />

Venkatraman and Ward choose slightly different ideas concerning information (Venkatraman) and<br />

enabling change (Ward). Whilst a detailed evaluation of the ideas would take more space than is<br />

available here, perusal of the table is revealing in that all perspectives align well, even though they see<br />

things from different perspectives. The merit of the IMBOK is that it is strongly aligned to the generation<br />

of value from an investment in information technology, and it is entirely aimed at fulfillment of the<br />

business strategy as the ultimate objective.<br />

Other threads of research have informed the development of the IMBOK. In its embryonic stages there<br />

was a very extensive review of the literature that established the validity of many of the ideas that have<br />

been used (Lambert & Peppard, 1993); the process value adding viewpoint was developed and published<br />

94


Andy Bytheway<br />

(Edwards & Peppard, 1997) and ideas about skills, competencies and capabilities were developed in<br />

conjunction with working business managers (Bytheway & Lambert, 1998). The concept of value came<br />

strongly into focus (Peppard et al., 2001) and is evident in standard works dealing with IS/IT strategy<br />

(Ward & Peppard, 2002).<br />

In South Africa, these ideas were incorporated into a research project that provided new learning material<br />

and new learning opportunities appropriate to the South African context (HICTE, 2003), and the IMBOK<br />

handbook was the result (Bytheway, 2004). It has since been adopted as a standard course text in South<br />

Africa, Europe and North America.<br />

3.1 The IMBOK framework<br />

Pictorially, the IMBOK identifies five domains of management, and four two-way interfaces between<br />

them.<br />

Information<br />

technology<br />

System<br />

projects<br />

System<br />

requirements<br />

Information<br />

system<br />

Business<br />

change<br />

Business<br />

needs<br />

Business<br />

process<br />

Business<br />

operations<br />

Business<br />

capability<br />

Business<br />

benefit<br />

Business<br />

results<br />

Goals and<br />

objectives<br />

Business<br />

strategy<br />

Figure 3: The Information Management Body of Knowledge<br />

It is clear that one of the principal difficulties faced by management is the preservation of the quality and<br />

detail of thinking that passes between the five management domains, and that there is not just one point<br />

of alignment (as is so often argued) but four. If information systems projects do not deliver systems that<br />

are well supported by the technology, if systems do not support business processes, if business<br />

processes do not deliver the performance improvements that are expected, and if those improvements<br />

are not what strategy demanded, then all will be in vain. Hence, the interfaces between the management<br />

domains take on special significance.<br />

The IMBOK framework is useful, because it allows assessment and analysis of the competencies that are<br />

needed to manage the successful delivery of benefits from information technology investments. From the<br />

preceding work, and in particular the review of a very wide range of literature (Lambert & Peppard, 1993),<br />

a set of 144 competencies have been identified that can be organized into nine groups that correspond to<br />

the five management domains in the IMBOK and the four gaps between them. These competencies are<br />

the substance of the survey instruments deployed in this work, and taken together they represent the<br />

capability of the organisation to achieve effective information management.<br />

3.2 The survey instruments<br />

The two survey instruments work at different levels:<br />

The first, with 144 statements each representing one of the 144 competencies; this is referred to as<br />

the "micro" level of working,<br />

The second, with a simplified form using just nine statements to represent each of the nine groups of<br />

competencies; this is referred to as the "macro" level of working.<br />

They are based on Likert-scale worksheets with statements, each representing one competency;<br />

respondents are asked to agree/disagree with the statements in order to indicate their perceptions of the<br />

competency of the organization to do things well. All statements are phrased in a positive sense, so that<br />

"agreement" is good news and "disagreement" is bad news. The focus on perceptions, as opposed to<br />

any absolute measure (such as might be found in using COBIT or ITIL) helps to take more careful<br />

account of people's feelings and emotions. Most previous thinking has revolved around structured,<br />

95


Andy Bytheway<br />

rational, right-brained thinking, but management is as much concerned with perceptions as with<br />

absolutes.<br />

3.3 The micro level<br />

Working at the micro level is of course potentially difficult: no respondent can be expected to respond to<br />

144 individual statements in what is hoped to be a quick but effective survey. Because a large number of<br />

respondents were expected to be involved, worksheets were prepared using randomly selected<br />

statements from the total set of 144, 20 statements at a time - enough worksheets would ensure that<br />

adequate coverage of the nine macro domains would be achieved.<br />

A fragment of such a worksheet is shown below:<br />

Figure 4: A sample worksheet from the survey instrument<br />

The statements are in no particular order and the significance of each statement (in terms of its position<br />

in the framework) is not revealed, so that they appear completely random to the respondent.<br />

Experience at the micro level soon indicated that at the start of an assessment there was a need for a<br />

short, simple but compatible approach. This was undertaken at the macro level.<br />

3.4 The macro level<br />

The simplified macro-level worksheet has just nine statements, one for each of the five domains and one<br />

for each of the four interfaces. Two additions were made: an additional statement concerning culture for<br />

change was included, and a rating of importance of each of the statements (high to low) was included so<br />

as to gather data about the relative importance of the different domains.<br />

Figure 5: The short survey instrument used at the macro level<br />

This short questionnaire proved to be an effective way of dealing with the hundreds of respondents<br />

involved with the financial services company.<br />

3.5 Four cases assessed<br />

The survey has been deployed in four cases, resulting (at the time of writing) in a total of almost 600<br />

responses and some thousands of individual opinions:<br />

A random sample of different businesses (and other organizations) in Cape Town, undertaken at the<br />

micro level.<br />

A purposeful sample of South African Chief Information Officers, also at the micro level.<br />

A purposeful sample of staff in all departments of a real estate agent, also at the micro level.<br />

96


Andy Bytheway<br />

A representative sample of staff working in a large financial services company, this time at the macro<br />

level.<br />

The results provide an interesting insight into the perceptions of different stakeholder groups: IT<br />

specialists, business specialists, administrative staff, management and others. For example, it is found<br />

that IT specialists often under-estimate their capability and fail to realise that their work is actually highly<br />

valued by others. In other cases senior management have a poor view of organisational ability to manage<br />

IT-specific issues such as technology acquisition, IT projects, and business change management.<br />

The paragraphs that follow present selected findings to illustrate these observations, and to show how<br />

the data is collated and presented back to the respondents using an analysis package that has itself<br />

developed over the period of the work.<br />

4. Results<br />

The opinions were analysed using simple, descriptive statistical methods, directed at the production of<br />

"radar charts". It has been found that radar charts are an effective way of presenting complex data to<br />

managers, and because the arrangement of summary measures around the circle is fixed, the audience<br />

for the results becomes used to the shapes that arise, indicating problems and opportunities.<br />

An analysis package has been developed, first to work with data at the micro level and later eanced to<br />

work at the macro level. It allows the selection of different datasets, sectors (of business), organizations<br />

within sector, and so on right down to the gender and education of the individual respondent, as will<br />

become evident. It automatically produces the radar charts (see below) that have ten spokes, or axes, as<br />

follows:<br />

IT The information technology management domain<br />

IT-IS The implementation of information systems<br />

IS The information systems management domain<br />

IS-BP The implementation of business change arising from new systems<br />

BP The business process management domain<br />

BP-BB The delivery of business benefits from improved business processes<br />

BB The business benefit management domain<br />

BB-BS The fulfillment of business strategy through performance<br />

BS The business strategy management domain<br />

(The five principal domains of the IMBOK are written in bold text, the four interfaces in italics).<br />

The tenth vertical "null" axis is reserved for "culture" measures - there was no data for that measure at<br />

this stage in the work.<br />

4.1 The reference sample<br />

The results for all received opinions (1391 in total) is shown in red (and bold); the results for the reference<br />

dataset (1076) are shown in blue (and feint); the standard deviation is shown in the centre in pale green.<br />

The nine measures (from IT through to BS, clockwise) all result in an average assessment of a little less<br />

than "3". This indicates a tendency on the part of respondents to respond to the statements using the<br />

third box from the right (they were coded "7" to "1", left to right):<br />

97


Figure 6: The results for the reference dataset<br />

Andy Bytheway<br />

Because the overall result is based on 1391 opinions, it might be expected that the results would "level<br />

out" (if perceptions of capability are indeed measured in an equitable way by the survey) - it is therefore<br />

pleasing that there is no excessive variation around the nine axes. Also, it can be seen that the standard<br />

deviation of data (on all axes) is in the range 1.3 to 1.5, which suggests that there are no areas of strong<br />

disagreement. It follows that deviations from these broadly-based figures will probably be significant.<br />

The figure below shows two selections that illustrate what is found in the detail:<br />

Figure 7: Results for "education" and "government" sectors selected from the reference dataset<br />

There are significant differences between these results: education is strong on strategy, but government<br />

is very weak; government considers its IT management to be strong but education does not. The<br />

perceived benefits to education are extremely poor, perhaps because of the strength of the strategic<br />

vision and the weakness of the management of the technology that should assist in its realization; the<br />

98


Andy Bytheway<br />

ability of government to see benefits coming out of routine business activity (in the business processes) is<br />

extremely poor.<br />

4.2 Chief Information Officers<br />

At a workshop in Cape Town, about 20 CIOs were asked to assist by completing worksheets at the micro<br />

level.<br />

Figure 8: Results for Chief Information Officers<br />

The results show a distinct optimism, especially when compared with some of the data from senior<br />

managers in the financial sector (see below). Perhaps this is not surprising, a CIO should feel optimistic<br />

about the merits of his work (but it would be a shock if this optimism was in distinct disagreement with the<br />

opinions of others in their organizations).<br />

4.3 Real estate agent<br />

The first in-company assessment concerned about 80 employees in a real estate agency, where the<br />

different views of the different groups proved to be interesting.<br />

Figure 9: Results from a real estate business<br />

Perhaps this is a typical result for a typical organisation, where overall there is a positive and optimistic<br />

view of information management (compared with the reference dataset). After all, this is an information<br />

intensive industry and it is critical to manage information well. This data was gathered at a time when<br />

there was a major upgrade taking place on the company web site, and therefore it is interesting to see<br />

more optimism in the IT department than in the Marketing department. When this analysis was made<br />

99


Andy Bytheway<br />

available to the management, it was possible to work to restore the balance. This is a marketing-led<br />

business, where it is essential that marketing staff share the general confidence in getting the technology<br />

to work well and deliver the benefits.<br />

4.4 Financial services company<br />

Because of the large number of potential respondents in this large financial services company, and<br />

because of the need to undertake the analysis quickly, this was the first application of the "simple"<br />

version of the survey instrument (the macro level). The analysis package being used was further<br />

developed, and so this data was more exhaustively analysed than before. The enhanced analyser allows<br />

for the selection of different departments and job levels, and it presents averages and standard<br />

deviations in detail for both capability and importance.<br />

Just two analyses are presented below: in the main image are the results of the whole set of 224<br />

responses (1906 opinions); overlaid at left-centre are the results for one senior manager who has quite<br />

different views than the overall average.<br />

The overall results show that there is broad concern about the level of capability to deal with IT, systems,<br />

and systems development. The results are less levelled than in the reference dataset. Given the large<br />

number of responses, this is a result that needs further examination. The senior manager has deep<br />

concerns about technology and systems capabilities (this is typical of other senior managers in this<br />

survey) and it is probably time to make sure that something is done to redress this concern.<br />

Having said all that, it is interesting that when compared with the reference dataset, the general results<br />

for this financial services business are good, averaging something between 4 and 5 (compared with less<br />

than 3 in the reference dataset). When presented with this data, the CIO of the financial services<br />

business was pleased, and indicated that more detailed results would be needed in order to guide<br />

management actions intended to improve information management capability. Work on this case<br />

continues.<br />

Figure 10: Analysis of macro data from a financial services business (overall result, with senior manager<br />

for comparison)<br />

100


5. Conclusions<br />

Andy Bytheway<br />

This paper has presented a review of the history of management thinking about information management<br />

issues, it has presented a framework that synthesizes the more interesting ideas, and has shown how<br />

this can be formalized as a new body of knowledge, a framework for thinking, and a capability<br />

assessment instrument.<br />

At the start of the paper, questions were asked that provide a basis for our conclusions:<br />

How can the unavoidable complexities be dealt with?<br />

How can costs be justified in the board room?<br />

Exactly what does the "alignment" of IT and business strategies mean?<br />

First, it is contended that the IMBOK deals with the complexities of information management in an<br />

effective and comprehensible way. It shows that there is no such thing as "alignment of IT and business<br />

strategies", it is necessary to align at four critical interfaces between five domains of management.<br />

Second, the IMBOK provides a means to communicate effectively with senior management. There are no<br />

technical terms, just a clear indication of the dependencies that justify the cost drivers (principally the<br />

acquisition of information technology and services) in light of the systems needed, the improvements to<br />

business processes, and the fulfillment of organizational strategy through appropriate benefits.<br />

Third, it is found that alignment needs a capability to effectively undertake all that is needed within the<br />

five domains of management, and at their interfaces. Space has precluded a detailed discussion of the<br />

144 competencies, but herein lies the secret to success: if any one of these many things is not done well,<br />

then the whole investment - value chain is at risk. No wonder that managing information well is difficult,<br />

and risky.<br />

It is still early days, but this research will continue. There must be more validation of the ideas and careful<br />

consideration of the merits of this kind of perceptions-based survey work. The analysis that is undertaken<br />

here is merely descriptive, and the potential for full inferential analysis of the data (statistically) remains to<br />

be examined.<br />

Acknowledgements<br />

This work would not have been possible without the enthusiastic involvement of staff and students at the<br />

University of the Western Cape and the Cape Technikon (now "CPUT"). Particular acknowledgement is<br />

due to the "ICT in Higher Education" research team led by Derek Keats, and to the Carnegie Corporation<br />

of New York that provided funding on that occasion.<br />

References<br />

Anon (2010) Labour's computer blunders cost £26,326 bn. The Independent. Available from:<br />

[Accessed 24 August 2010].<br />

Anon (2008) Making Sure Tech Is in Step. Wall Street Journal - Eastern Edition, 252 (59), p.B9.<br />

Baets, W. (1992) Aligning information systems with business strategy. The Journal of Strategic Information Systems,<br />

1 (4), pp.205-213.<br />

Beaudry, A. & Pinsonneault, A. (2010) The other side of acceptance: Studying the direct and indirect effects of<br />

emotions on information technology use. MIS Quarterly, 34 (4), pp.689-A3.<br />

Bytheway, A. (1996) Strategic information modelling: some experiences with a new framework. Journal of Applied<br />

Management Studies, 5 (1), pp.17-45.<br />

Bytheway, A. (2004) The Information Management Body of Knowledge. 1st ed. World Wide Web. Available from:<br />

.<br />

Bytheway, A. & Braganza, A. (1997) Process based collaboration: strategies, tactics and benefits. Cranfield UK,<br />

Cranfield School of Management.<br />

Bytheway, A. & Lambert, R. (1998) Organisational competencies for harnessing IS/IT. Available from:<br />

[Accessed 7 May 2010].<br />

Chan, Y.E. & Reich, B.H. (2007) IT alignment: what have we learned? Journal of Information Technology, 22 (4),<br />

pp.297-315.<br />

Chatterji, S. (2007) Bridging business and IT strategies with Enterprise Architecture: Realising the real value of<br />

business-IT alignment. Information Systems Control Journal|, 3, pp.1-2.<br />

Chen, D.Q., Mocker, M., Preston, D.S. & Teubner, A. (2010) Information Systems strategy: Reconceptualization,<br />

measurement and implications. MIS Quarterly, 34 (2), pp.233-A8.<br />

Cross, J. (1995) IT outsourcing : British Petroleum's competitive approach. Harvard Business Review, 73 (3), p.94.<br />

Cross, J. & Earl, M. (1997) Transformation of the IT function at British Petroleum. MIS Quarterly, 21 (4), pp.401-424.<br />

101


Andy Bytheway<br />

DeLone, W. & McLean, E. (1992) Information Systems Success: The Quest for the Dependent Variable. Information<br />

Systems Research, 3 (1), pp.60-95.<br />

Edwards, C. & Peppard, J. (1997) Operationalizing strategy through process. Long Range Planning, 30 (5), pp.753-<br />

767.<br />

Frankel, D.S., Harmon, P., Mukerji, J., Odell, J., Owen, M., Rivett, P., Rosen, M. & Soley, R.M. (2003) The Zachman<br />

Framework and the OMG's model driven architecture. Business Process Trends (Whitepaper). Available from:<br />

[Accessed 22<br />

November 2010].<br />

Henderson, J.C. & Venkatraman, N. (1993) Strategic alignment: leveraging information technology for transforming<br />

organizations. IBM Systems Journal, 32 (1), pp.4-16.<br />

HICTE (2003) ICT in Higher Education. Cape Town, University of the Western Cape.<br />

Kearns, G.S. & Lederer, A.L. (2000) The effect of strategic alignment on the use of IS-based resources for<br />

competitive advantage. The Journal of Strategic Information Systems, 9 (4), pp.265-293.<br />

Lambert, R. & Peppard, J. (1993) IT and new organizational forms: destination but no road map? Journal of Strategic<br />

Information Systems, 2 (3), pp.180-206.<br />

Mingers, J. & Walsham, G. (2010) Toward ethical information systems: The contribution of discourse ethics. MIS<br />

Quarterly, 34 (4), pp.833-854.<br />

Peppard, J., Lambert, R. & Edwards, C. (2001) Whose job is it anyway? Information Systems Journal, 10 (4), pp.291-<br />

322.<br />

Sambamurthy, V., Bharadwaj, A. & Grover, V. (2003) Shaping agility through digital options: Reconceptualizing the<br />

role of IT in contemporary firms. MIS Quarterly, 27 (2), pp.237-264.<br />

Seddon, P.B., Calvert, C. & Yang, S. (2010) A mulit-project model of key factors affecting organizational benefits<br />

from Enterprise Systems. MIS Quarterly, 34 (2), pp.305-A11.<br />

Symons, C. (2005) IT Governance Framework. Forrester. Available from:<br />

[Accessed 22 November 2010].<br />

Tallon, P.P., Kraemer, K.L. & Gurbaxani, V. (2000) Executives' perceptions of the business value of information<br />

technology: a process-oriented approach. Journal of Management Information Systems, 16 (4), pp.145-173.<br />

Uchitelle, L. (1996) What has the computer done for us lately? New York Times, p.21.<br />

Venkatraman, N. (1994) IT-enabled business transformation: from automation to business scope redefinition. Sloan<br />

Management Review, 35 (2), pp.73-87.<br />

Venkatraman, N. (1996) Managing IT resources as a value center. IS Executive Seminar Series, Cranfield School of<br />

Management.<br />

Ward, J. & Daniel, E. (2005) Benefits Management: Delivering Value from IS and IT Investments. Chichester, Wiley.<br />

Ward, J. & Peppard, J. (2002) Strategic Planning for Information Systems (3rd Edition). Chichester, Wiley.<br />

Yayla, A. & Hu, Q. (2009) Antecedents and drivers of IT-Business Strategic Alignment: Empirical validation of a<br />

theoretical model. In: Proc 17th <strong>European</strong> <strong>Conference</strong> on Information Systems. Verona Italy.<br />

Zachman, J.A. (1987) A Framework for Information Systems Architecture. IBM Systems Journal, 26 (3), pp.590-616.<br />

102


A Semiotic Approach to Analyse the Influencing Factors in<br />

Knowledge Transfer<br />

Supaporn Chai-Arayalert 1 and Keiichi Nakata 2<br />

Henley Business School, University of Reading, UK<br />

s.chai-arayalert@pgr.reading.ac.uk<br />

k.nakata@henley.reading.ac.uk<br />

Abstract: With the rapid growth of information and technology, knowledge is a valuable asset in organisation which<br />

has become significant as a strategic resource. Many studies have focused on managing knowledge in<br />

organisations. In particular, knowledge transfer has become a significant issue concerned with the movement of<br />

knowledge across organisational boundaries. It enables the exploitation and application of existing knowledge for<br />

other organisations, reducing the time of creating knowledge, and minimising the cost of organisational learning. In<br />

order to achieve their goals and objectives, organisations need to combine knowledge and performance<br />

measurement tools such as Key Performance Indicators (KPIs) which are powerful agents of organisational change<br />

that help organisations to measure their progress towards achieving their goals. In this paper, we discuss how<br />

organisations can transfer knowledge through KPIs effectively by identifying the relationship between knowledge<br />

transfer and KPIs. In addition, we analyse localisation factors of knowledge transfer based on a semiotic approach:<br />

the organisational containment analysis to scope the influencing localisation factors, and the semiotic framework to<br />

analyse the socio-technical aspects of these factors.<br />

Keywords: knowledge management, knowledge transfer, key performance indicator, semiotics<br />

1. Introduction<br />

Knowledge is a valuable asset in organisations and it has become significant as a strategic resource.<br />

Many studies have focused on managing knowledge in organisations with various processes and<br />

methods (Alavi and Leidner 2001). Knowledge transfer has recently become a significant issue in<br />

knowledge management concerned with the movement of knowledge across the boundaries (Von Krogh<br />

and Roos 1996, Cummings 2003). In addition, it enables the exploitation and application of existing<br />

knowledge for an organisation’s purposes. In order to achieve its goals and objectives, the organisation<br />

needs to combine knowledge and performance measurement tool such as Key Performance Indicators<br />

(KPIs). These are powerful agents of organisational change that help organisations to measure their<br />

progress towards achieving their goals. One of the most important reasons for measuring performance is<br />

to assess the success rate of the business as well as its current performance. KPIs are often used to<br />

help an organisation to achieve its long term organisational goals. Therefore, it is useful to analyse how<br />

performance indicators represent knowledge which in turn could be used for knowledge transfer. In this<br />

paper, we discuss how organisations can transfer KPIs as knowledge effectively by identifying the<br />

relationship between knowledge transfer and KPIs. In addition, we analyse localisation factors in<br />

knowledge transfer based on the semiotic approach. This paper is organised as follows. First, the<br />

concept of knowledge management and knowledge transfer is discussed, followed by a review of the<br />

concepts of KPIs and organisational semiotics. Then, methods to analyse the localisation factors of<br />

knowledge transfer based on semiotics are proposed and followed by the conclusion.<br />

2. Background<br />

In this section, we review knowledge management and its activities and the notion of knowledge transfer,<br />

followed by an overview of key performance indicators and their relation to knowledge.<br />

2.1 Knowledge management<br />

In recent years, there has been a transformation from the resource based view of an organisation into the<br />

knowledge based view (Alavi and Leidner 2001). As a result, more organisations are treating knowledge<br />

as a vital organisational resource alongside other assets, which has led the organisations to recognise<br />

the importance of knowledge management.<br />

In the organisational context, there are many definitions of knowledge which reflects the multiple ways in<br />

which knowledge is considered to be used within organisations. Nonaka and Takeuchi (1995) define<br />

knowledge as a dynamic human process of justifying personal belief toward the truth. Prusak and<br />

Davenport (1998) define knowledge to include experiences, beliefs, values, how we feel, motivation and<br />

information. Additionally, they focuses on the function or purpose of knowledge which is a framework that<br />

people use for evaluating and incorporating new experiences and information through embedded<br />

103


Supaporn Chai-Arayalert and Keiichi Nakata<br />

routines, processes, practices, and norms. Arthur Anderson and APQC (1996) define knowledge as the<br />

relationship between knowledge and organisation as well as information that has value and the collective<br />

experience of the organisation. In the same vein, Demarest (1997) defines knowledge as the actionable<br />

information embodied in work practices, theories-in-action, skills, equipment, processes and heuristics of<br />

firm's employees.<br />

Knowledge management (KM) is to identify and leverage the collective knowledge in an organisation to<br />

help the organisation to be competitive (Von Krogh and Roos 1996). Activities in KM consist of four basic<br />

processes of creating, storing or retrieving, transferring, and applying knowledge (Von Krogh and Roos<br />

1996). Knowledge creation means developing new contents or replacing existing knowledge within<br />

organisational knowledge. Knowledge storage or retrieval refers to organisational memory involving<br />

documentation, database, and codified human knowledge stored in expert systems, etc. Moreover, an<br />

important process in KM is knowledge transfer. Knowledge transfer is the re-creation of a source’s<br />

knowledge-related elements in the recipient (Cummings 2003). Knowledge application is a vital aspect of<br />

knowledge-based theory of the firm, which is the source of competitive advantage. This process is<br />

considered as an activity to apply knowledge to different places and learning from it (Bouthillier and<br />

Shearer 2002).<br />

2.2 Knowledge transfer<br />

Knowledge transfer has become one of the significant KM processes concerned with the movement of<br />

knowledge across the boundaries created by specialised knowledge domains (Carlile 2002, Carlile and<br />

Rebentisch 2003). It is the movement of knowledge from one place, person or ownership to another.<br />

Furthermore, knowledge transfer enables the exploitation and application of existing knowledge for the<br />

organisation’s purposes.<br />

The concept of knowledge transfer based on the communication theory is defined as “a process of<br />

exchange of explicit or tacit knowledge between two agents, during which one agent purposefully<br />

receives and uses the knowledge perceived by another” (Kumar and Ganesh 2009 p.163). Moreover, the<br />

knowledge transfer is “identifying (accessible) knowledge that already exists, acquiring it and<br />

subsequently applying this knowledge to develop new ideas or enhance the existing ideas to make a<br />

process/action faster, better or safer than they would have otherwise been” (Liyanage et al. 2009, p.122).<br />

This suggests that in knowledge transfer, consideration should also be given to the acquisition of<br />

knowledge as well as how it can be exploited to improve organisational activities.<br />

Successful knowledge transfer can be explained as the results of transfer in the receiving end<br />

accumulating or assimilating new knowledge. In other words, success of knowledge transfer can be<br />

assessed by how the underlying knowledge elements have been re-created in the recipient to conform to<br />

those of the source (Cummings 2003). While the barriers of knowledge transfer from source,<br />

communication channel and recipient have been studied (Khamseh and Jolly 2008, McLaughlin 2007,<br />

Jensen and Szulanski 2004), previous studies have not analysed both human and technical aspects<br />

together.<br />

2.3 Key performance indicators and their relation to knowledge<br />

Key performance indicators (KPIs) measure organisation’s overall health and well being (AusIndustry<br />

1995). Importantly, it focuses on the areas of organisation’s performance which are critical for ongoing<br />

and future success. KPIs are driver tools to change an organisation and its culture; at the same time,<br />

organisations can apply KPIs as an improvement strategy where they will need to change (Swan and<br />

Kyng 2004). KPIs are commonly used to help an organisation define and evaluate its success, in terms of<br />

making progress in long term organisational goals and toward knowledge-based goals which are difficult<br />

to quantify. The differences in KPIs depend on the nature and characteristics of the organisation including<br />

its strategies and goals. This section examines the concept of KPIs and the relationship between KPIs<br />

and organisational knowledge.<br />

2.3.1 The concept of key performance indicators<br />

KPIs are often defined as the quantifiable metric which reflects the organisational performance in<br />

achieving its goals and objectives, which in turn it reflects strategic drivers and critical success factors of<br />

organisational activities and processes (Parmenter 2007, Thanyaphut 2006). Furthermore, KPIs are<br />

indicators at the operational and tactical level obtained by breaking down strategic objective decision<br />

104


Supaporn Chai-Arayalert and Keiichi Nakata<br />

making and act as pointers for monitoring the effect of implementation of macro strategic decision (Zhou<br />

et al. 2006). KPIs can align to units, departments and individuals of an organisation. An organisation can<br />

have one or more KPIs to measure its outcomes. KPIs have to be created based on the organisational<br />

vision, mission, and strategy. Therefore, KPIs usually differ in each organisation due to the different goals<br />

and strategies. Organisations can utilise KPIs to monitor progress, communicate performance results,<br />

and determine priorities (Promput et al. 2006). Organisations can be monitored by setting KPIs, and KPIs<br />

are also created to identify the organisational position compared to its competitors and to communicate<br />

performance results within the organisations. KPIs show goal achievement results based on facts, not on<br />

emotion or supposition so that employees will accept and understand the results. To determine work<br />

priorities, KPIs show the target employees need to achieve and helps them set the priorities of their work.<br />

2.3.2 Organisational knowledge and key performance indicators<br />

As described earlier, knowledge management (KM) is recognised as a part of strategy to improve<br />

business performance. KPIs complement performance measurement and they are a part of<br />

organisational knowledge which is most vital in improving performance. Therefore, it has been suggested<br />

that managing knowledge could be incorporated into KPIs, and other performance measurement<br />

approaches (Okkonen 2004). KM consists of activities and tools which are used to improve core<br />

competencies and to make the use of information more efficiently in organisational context (Okkonen<br />

2004). A relationship between KM and performance measurement has also been identified (Okkonen et<br />

al. 2002), i.e., performance measures provide the information needs in KM activities. In addition, the<br />

process of formulating and implementing measures or indicators may be through gathering and<br />

implementing organisational knowledge. As KPIs depend on objectives or goals of organisational<br />

overview and break down strategic knowledge into lower levels or operational knowledge in organisation,<br />

they can be considered as explicit knowledge due to the fact that they are captured as written<br />

documents, forms, tables etc. (Promput et al. 2006).<br />

There are a number of limitations of developing and using KPIs. Firstly, the development KPIs is specific<br />

to an organisation and uses information from different sources. The relevant KPIs can be found in internal<br />

or external documents within the organisation, e.g., policies, mission statements, business plan, job<br />

descriptions, and laws. Therefore, acquisition of knowledge of KPIs development may be incomplete and<br />

vague (Popova and Sharpanskykh, 2010). Secondly, KPIs can be developed by using organisational<br />

knowledge, time, and needs experts to facilitate the KPI development process. However, some<br />

organisations are not conveniently set up to capture knowledge, find it time consuming, and sometimes<br />

lack expertise (Popova and Sharpanskykh, 2010). Thirdly, although KPIs are useful in performance<br />

measurement due to the fact that the organisation’s objectives are broken down into KPIs that are<br />

applied to measure some aspects against targets in order to improve the problems areas, the KPI<br />

implementation can also suffer from the same reasons performance measurement initiatives fail, as<br />

argued by Neely and Bourne (2000) and Masayna et al (2007). They state that these initiatives fail<br />

because they do not reflect the strategies and help people to understand what the organisation’s priorities<br />

are. Another reason is the lack of infrastructure which is focused on the management of data in<br />

developing and using KPIs (Neely and Bourne, 2000). This problem of data management in the<br />

organisation is that they are often unrelated, unlinked, and inconsistent (Masayna et al., 2007).<br />

3. Analysing localisation factors of knowledge transfer<br />

This section discusses the relationship between KPIs and organisational knowledge, the concept of<br />

knowledge transfer through KPIs, and factors that influence this form of knowledge transfer. It begins by<br />

describing organisational semiotics which is applied for problem analysis, especially the semiosis model,<br />

the organisational containment analysis, and the semiotic framework. This is followed by an exploration<br />

of localisation factors based on the semiosis model. Finally, we analyse these localisation factors in<br />

knowledge transfer using organisational containment analysis and semiotic framework.<br />

3.1 Organisational semiotics<br />

Semiotics presents signs as things in a world of objects, actions and relationships. As a branch of<br />

semiotics, organisational semiotics (OS) is a discipline which aims to study the nature, functions,<br />

characteristics and effects of information and communication within organisational contexts (Liu 2000).<br />

OS helps us understand the cooperative workings and interactions among individuals, and between<br />

human beings and technology. It defines organisations as systems where signs are created and used for<br />

communication and business purposes (Liu et al. 1999). It deals with the use of signs and the<br />

construction of shared meanings within and among organisations (Liu 2000).<br />

105


3.2 Peirce’s triadic model of semiosis<br />

Supaporn Chai-Arayalert and Keiichi Nakata<br />

Semiosis, introduced by Charles Sanders Peirce (1839–1914), is the process of constructing meaning<br />

from represented signs. The process is shown by Peirce’s Triadic Model of semiosis. Semiosis contains<br />

sign, object and interpretant (Liu 2000). Sign is the signification without reference to anything other than<br />

itself. The object is the signification in relation to something else. The interpretant is shown as to meditate<br />

the relationship and helps establish the mapping between the sign and the object (presented by the<br />

dotted line in figure 1). The sign is related to its referent or the object with the assistance of the<br />

interpretant which is the interpretation process (from sign to object). The sign can be understood or<br />

misunderstood in different ways depending on the interpretant. The semiosis model can assist the<br />

analysis of knowledge transfer, as the interactions between the sign, object and interpretant.<br />

Figure 1: Peirce’s triadic model of semiosis (Liu 2000)<br />

Based on this view, we can analyse semiosis gaps in knowledge transfer processes. There are two<br />

possible gaps: representation gaps and interpretation gaps (Tan et al. 2003). ‘Representation’ is to<br />

describe something or illustration of a sign. A representation gap occurs when the two corresponding<br />

signs referring to the same object are not aligned. Secondly, an interpretation gap is the displacement of<br />

objects or description leading to a mismatch of inference. This gap happens when understanding of<br />

object differs and results in a distorted understanding of intended meaning. Employing the semiosis<br />

model can explain gaps that might occur in knowledge transfer.<br />

The figure 2 illustrates the notion of semiosis gaps.<br />

Figure 2: The semiosis gaps<br />

106


Supaporn Chai-Arayalert and Keiichi Nakata<br />

The semiosis model presents two significant gaps occurring in the sign (S) and the interpretant (I). The<br />

sign represents KPIs which an organisation intends to use as a vehicle to transfer knowledge to another<br />

organisation. In this model, S1 represents developed KPIs from source’s organisation and S2 occurs when<br />

S1 is adapted through adaptation procedures that are concerned with localisation barriers of transferring<br />

KPIs. The object (O) is knowledge. In the interpretant (I) is the process of relating KPIs to organisational<br />

knowledge. We adopt the view that knowledge transfer can be carried out through adapting KPIs of one<br />

organisation to another, the transfer of knowledge to external organisation and different countries can be<br />

influenced by factors which affect the success of transfer. These factors are represented by gaps<br />

between the interpretant (I1) and interpretant (I2). In the next section, we use the semiotic framework to<br />

identify the vital localisation factors.<br />

3.3 Analysis of influencing localisation factors based on semiotic approach<br />

This analysis of localisation factors that influence knowledge transfer begins with collecting data from<br />

literature, after which the data is scoped by organisation containment analysis and analysed by the<br />

semiotic framework.<br />

3.3.1 Scoping the influencing localisation factors based on organisational containment analysis<br />

The organisational containment analysis (figure 3) presents a model of organisation as a system in terms<br />

of informal, formal and technical layers (Liu 2000). The informal layer covers the whole business<br />

organisation and it interacts directly with the external environment and context. It focuses on subcultures,<br />

behaviours, beliefs and practices. The activities of this layer can be formalised in the formal<br />

layer or as business rules. The formal layer consists of forms, rules, and procedures. These formalisms<br />

are developed to clarify and replace meanings, intentions, commitments, and responsibilities of the<br />

informal layer. In the technical layer, technical systems are designed to automate parts of formalised<br />

aspects, which in turn rely on the informal aspect.<br />

Figure 3: An organisational containment analysis for scoping localisation factors<br />

Here we attempt to identify the localisation factors that influence knowledge transfer. The informal layer<br />

includes the national culture and norms (Liyanage et al. 2009, Kluge et al., 200, Barson et al. 2000); the<br />

experience of source and recipient organisation (Liyanage et al. 2009, Khamseh and Jolly 2008, Jensen<br />

and Szulanski 2004); the ability to identify, assimilate, transform and apply external knowledge or<br />

absorptive capacity (Khamseh and Jolly 2008); and language barriers (Cranefield and Yoong 2007,<br />

Jensen and Szulanski 2004). In the formal layer, the factors related to the characteristics of target<br />

organisations including strategy and objectives (AusIndustry 1995), organisational structure (Cranefield<br />

107


Supaporn Chai-Arayalert and Keiichi Nakata<br />

and Yoong 2007, McLaughlin 2007, Barson et al. 2000, AusIndustry 1995), technical terms (Cranefield<br />

and Yoong 2007, McLaughlin 2007), regulations and rules (McLaughlin 2007), standards (McLaughlin<br />

2007), policies (AusIndustry 1995). As described above, the technical layer refers to specific factors of<br />

information technology involving hardware, software and communication channels (Barson et al. 2000,<br />

Wathne et al. 1996, Khamseh and Jolly 2008, Cranefield and Yoong 2007).<br />

3.3.2 Analysing the influencing localisation factors based on the semiotic framework<br />

Based on the scoping of localisation factors above, we apply the semiotic framework to analyse the<br />

localisation factors which influence the transfer of knowledge through KPIs. The semiotic framework was<br />

developed in order to understand different dimensions of signs (Stamper 1996). This framework is<br />

divided as human and technical aspects. The technical aspects are the physical world, the empiric layer<br />

and the syntactic layer. The upper three layers address the human aspects in terms of the semantic<br />

layer, the pragmatic layer and the social world (table 1).<br />

Table 1: The semiotic framework for analysing localisation factors of knowledge transfer<br />

The physical world is concerned with the physical aspects of signs at the level of signals and marks (Liu<br />

2000). The localisation factors in this layer focus on channel infrastructures, technologies and resources.<br />

The channel infrastructures can facilitate the knowledge transfer process (Wathne et al. 1996). However,<br />

there are barriers to transfer in terms of digital divide in some countries. Therefore, some countries<br />

cannot conveniently transfer and adapt knowledge from another organisation. In addition, the barriers of<br />

inter-and-intra knowledge transfer are technologies in term of existing resource, available technology and<br />

legacy systems (Barson et al., 2000).<br />

The empiric layer is statistical properties of signs when different physical media and devices are used<br />

(Liu 2000). This layer is concerned with the statistical behaviour of signs and it can be matched to the<br />

statistical characteristic of the media. In empirics view, information is viewed as a stream of signals which<br />

is transmitted from sender to receiver. Therefore, the analysis of localisation factors is related to video-<br />

108


Supaporn Chai-Arayalert and Keiichi Nakata<br />

conference, telephone, internet, documents, and face-to-face meetings(Cranefield and Yoong 2007)<br />

which have an effect on transferring knowledge.<br />

The syntactic layer is concerned with the complex structure of a language (Liu 2000). Study of this layer<br />

will answer question as to how signs are structured and used in language. We identify the localisation<br />

factors by looking at the characteristics of knowledge, i.e., metrics, variables, and formula (McLaughlin<br />

2007, Khamseh and Jolly 2008, Zhihong 2008). In addition, the procedure and/or process, software and<br />

database technology are considered as localisation factors which affect both the source and the recipient<br />

(Khamseh and Jolly 2008).<br />

The semantic layer is considered with the relationship between a sign and what it refers to (Liu 2000).<br />

The transfer and adaptation of knowledge through KPIs between organisations of different backgrounds<br />

need to share the same understanding and meaning of knowledge. The knowledge transfer from the<br />

source to the recipient cannot take place, if they do not understand and share the meanings when<br />

adopting and adapting KPIs. The localisation factors in this layer are technical terms, specific definitions,<br />

standards and language (Cranefield and Yoong 2007).<br />

The pragmatic layer focuses on the relationships between signs and behaviours of agents (Liu 2000).<br />

This layer considers signs that can be used for certain purposes or objectives. In the case of transferring<br />

knowledge through KPIs to difference countries, although organisations may have a similar domain, they<br />

may differ in their context. Therefore, not all indicators can be transferred and applied as objectives and<br />

strategies in another organisation. It depends on the organisational context such as types of organisation,<br />

organisational structure, strategies, and objectives which directly influence the behaviours of adapting<br />

indicators (AusIndustry 1995, Cranefield and Yoong 2007, McLaughlin 2007, Jensen and Szulanski<br />

2004). The difference in organisational context can make it difficult to transfer knowledge between<br />

organisations.<br />

The social world is the effect of the use of signs in social and human affairs (Liu 2000). In particular, it<br />

implicates social norms, beliefs, expectations, commitments, functions, contracts, law, and culture (Filipe<br />

and Liu 2000). KPIs can be used to measure performance in organisations to achieve objectives and<br />

goals. They are applied to measure the performance against targets which are intended to solve or<br />

improve areas of weakness. Knowledge can be embedded in KPIs which can be considered as<br />

representation of knowledge which is generated from several sources. Formalising KPIs requires tangible<br />

and intangible resources. Tangible resource includes reports, manuals, stories, and records etc. and<br />

intangible resource are culture, experience, norms, and commitments. The knowledge transfer through<br />

KPIs occurs in the social world layer; however, this is not a trivial process for the limitation of KPIs<br />

development mentioned above (Alwaer and Clements-Croome 2010). The localisation factors can be<br />

captured as the interpretant gap in the semiosis model, and includes national culture, laws and<br />

regulations, experiences, and absorptive capacity. National culture constrains knowledge adaptation as<br />

well as governs the process of transfer (Davenport and Prusak 1998). As such, national culture is clearly<br />

a significant and complex issue in knowledge transfer. Others suggest that knowledge transfer process<br />

cannot succeed if the source and recipient are unwilling to share knowledge because of cultural<br />

difficulties (Liyanage et al. 2009). Experience is found as one of the important factors. A number of<br />

studies analysed experience as a key factor impacting on knowledge transfer. Cranefield and Yoong<br />

(2007) identified the key factors that impacted on knowledge transfer and the result showed the prior<br />

experience had an impact on knowledge transfer during the first stage of transfer. Absorptive capacity is<br />

the ability to identify, assimilate, transform, and apply valuable external knowledge (Cohen and Levinthal<br />

1990). Therefore, the recipient may be unable to utilise external sources of knowledge because they may<br />

lack absorptive capacity which affects their ability to value, assimilate and apply new knowledge<br />

(Szulanski 1996).<br />

4. Conclusion<br />

An effective acquisition and management of knowledge has become a competitive advantage in the use<br />

of organisational resources. The organisational knowledge can be captured through learning, which<br />

includes the transfer of knowledge among organisations. However, the transfer of knowledge is not<br />

straightforward as it depends not just on the nature of knowledge itself but also on the process of<br />

acquiring and assimilating it. This study explored methods of analysis and identification of localisation<br />

factors. It aimed to identify the relationships between knowledge transfer and its vehicles such as KPIs<br />

through the use of semiotic approach. The localisation factors are also identified using the semiotic<br />

approach. The outcomes of this paper are as follows. Firstly, we used organisational containment<br />

109


Supaporn Chai-Arayalert and Keiichi Nakata<br />

analysis, which analyses an organisation as a system made up of informal, formal and technical layers, to<br />

scope the influencing localisation factors when knowledge transfer is carried out through KPIs. As a<br />

result, the localisation factors with the informal layer have been identified as the culture and norms; the<br />

experiences of source and recipient organisation; the ability to identify, assimilate, transform and apply<br />

external knowledge or absorptive capacity; and language barriers. In the formal layer, the factors are<br />

related to the characteristics of target organisations including strategy, objectives, organisational<br />

structure, technical terms, definitions, language, rules, standards, policies, and laws. The technical layer<br />

refers to specific factors of information technology involving hardware, software and communication<br />

infrastructure. Secondly, the influencing localisation factors were divided into human and technical<br />

aspects based on the semiotic framework. In the future work, we plan extend to this research by<br />

developing a conceptual framework of knowledge transfer based on the result of this analysis, and<br />

validate it through case studies in the inter-organisational and inter-cultural knowledge transfer.<br />

References<br />

Alavi, M. and Leidner, D. E. (2001) 'Review: Knowledge management and knowledge management systems:<br />

Conceptual foundations and research issues', MIS quarterly, pp.107-136.<br />

AusIndustry (1995) Key performance indicators manual : a practical guide for the best practice development,<br />

implementation and use of KPIs., Business & Professional Publishing..<br />

APQC. (1996) The Knowledge Management Assement Tool, Arthur Anderson and The American Productivity &<br />

Quality Center.<br />

Alwaer, H. and Clements-Croome, D. J. (2010) 'Key performance indicators (KPIs) and priority setting in using the<br />

multi-attribute approach for assessing sustainable intelligent buildings', Building and Environment, 45(4), 799-<br />

807.<br />

Barson, R. J., Foster, G., Struck, T., Ratchev, S., Pawar, K., Weber, F. and Wunram, M. (2000) Inter-and intra-<br />

Organisational barriers to sharing knowledge in the extended supply-chain,IOS Press.<br />

Bouthillier, F. and Shearer, K. (2002) 'Understanding knowledge management and information management: the<br />

need for an empirical perspective', Information research, 8(1).<br />

Carlile, P. (2002) 'Transferring, Translating and Transforming: An Integrative Framework Managing Knowledge<br />

across Boundaries', Effective Enterprise Learning, pp. 555-568.<br />

Carlile, P. R. and Rebentisch, E. S. (2003) 'Into the black box: the knowledge transformation cycle', Management<br />

Science, 49(9), pp.1180-1195.<br />

Carrillo, P. M., Robinson, H. S., Anumba, C. J. and Al-Ghassani, A. M. (2003) 'IMPaKT: A framework for linking<br />

knowledge management to business performance', Electronic Journal of Knowledge Management, 1(1), pp.1-<br />

12.<br />

Cohen, W. M. and Levinthal, D. A. (1990) 'Absorptive capacity: a new perspective on learning and innovation',<br />

Administrative science quarterly, 35(1), pp 128-152.<br />

Cranefield, J. and Yoong, P. (2007) 'Organisational factors affecting inter-organisational knowledge transfer in the<br />

new zealand state sector: a case study', Journal of Organizational Virtualness, 1, pp. 9-33.<br />

Cummings, J. (2003) Knowledge Sharing: a review of the literature, The world Bank Operations Evaluation<br />

Department, The World Bank, Washington D.C.<br />

Davenport, D. L. and Hplapple, C. W. (2006) 'Knowledge organizations' in Schwartz, D., ed. Encyclopedia of<br />

knowledge management,Idea Group Reference, pp.451-458.<br />

Davenport, T. and Prusak, L. (1998) Working knowledge: How organizations manage what they know, Harvard<br />

Business School.<br />

Demarest, M. (1997) 'Understanding knowledge management', Long Range Planning, 30(3), pp. 374-384<br />

Filipe, J. and Liu, K. (2000) The EDA Model: An Organizational Semiotics Perspective To Norm-Based Agent Design,<br />

Workshop on Norms and Institutions in Multi-agent systems at Normative agent.<br />

Jensen, R. and Szulanski, G. (2004) 'Stickiness and the Adaptation of Organizational Practices in Cross-Border<br />

Knowledge Transfers', Journal of International Business Studies, 35(6), pp.508-524.<br />

Khamseh, H. M. and Jolly, D. R. (2008) 'Knowledge transfer in alliances: determinant factors', Journal of Knowledge<br />

Management, 12(1), pp.37-50.<br />

Kluge, J., W. Stein, and T. Licht. (2001) ‘Knowledge unplugged: The Mckinsey & Company global survey on<br />

knowledge management’, Palgrave Macmillan.<br />

Kuhn, O. and Abecker, A. (1997) 'Corporate memories for knowledge management in industrial practice: Prospects<br />

and challenges',Journal of Universal Computer Science, 3(8), pp.929-954.<br />

Kumar, J. A. and Ganesh, L. S. (2009) 'Research on knowledge transfer in organizations: a morphology', Journal of<br />

Knowledge Managemnt, 13(4), pp.161-174.<br />

Liu, K. (2000) Semiotics in information systems engineering, Cambridge:Cambridge University Press.<br />

Liu, K., Alderson, A., Shah, H., Sharp, B. and Dix, A. (1999) 'Applying semiotic methods to requirements recovery',<br />

Methodologies for Developing and Managing Emerging Technology-Based Information Systems, pp.142-152.<br />

Liyanage, C., Elhag, T., Ballal, T. and Li, Q. (2009) 'Knowledge communication and translation–a knowledge transfer<br />

model', Journal of Knowledge Management, 13(3), 118-131.<br />

Masayna, V., Koronios, A., Gao, J. and Gendron, M. (2007) 'Data quality and KPIs: a link to be established', in The<br />

2nd World Congress on Engineering Asset Management (EAM) and The 4th International <strong>Conference</strong> on<br />

Condition Monitoring,<br />

110


Supaporn Chai-Arayalert and Keiichi Nakata<br />

McLaughlin, S. (2007) Identifying knowledge transfer barriers within a complex supply chain organization.,<br />

unpublished Ph.D.thesis University of Glasgow<br />

Neely, A. and Bourne, M. (2000) 'Why measurement initiatives fail', Measuring Business Excellence, 4(4), 3-7.<br />

Nonaka, I. and Takeuchi, H. (1995) 'The knowledge-creating company: How Japanese companies create the<br />

dynamics of innovation'<br />

Okkonen, J. (2004) The Use of Performance Measurement in Knowledge Work Context, unpublished Ph.D.thesis<br />

University of Tampere.<br />

Okkonen, J., Pirttim ki, V., Hannula, M. and L nnqvist, A. (2002) 'Triangle of Business Intelligence, Performance<br />

Measurement and Knowledge Management', in 2nd Annual <strong>Conference</strong> on Innovative Research in<br />

Management, Stockhome, Sweden, 2002, pp.1-14<br />

Parmenter, D. (2007) Key performance indicators: developing, implementing, and using winning KPIs, Wiley.<br />

Popova, V. and Sharpanskykh, A. (2010) 'Modeling organizational performance indicators', Information Systems,<br />

35(4), 505-527.<br />

Promput, W., Pingsuttiwong, S. and Lerdmahakit, C. (2006) KPIs principle to practice, Bangkok:Siriwattana Interprint.<br />

Stamper, R. (1996) 'Signs, Norms and Information Systems, Signs at Work' in al, Holmqvist B. et al, ed.,Berlin:<br />

Walter de Gruyter.<br />

Swan, W. and Kyng, E. (2004) An Introduction to Key Performance Indicators, Manchester, UK:Centre for<br />

Construction Innovation.<br />

Szulanski, G. (1996) 'Exploring internal stickiness: Impediments to the transfer of best practice within the firm',<br />

Strategic management journal, 17(1), pp.27-43.<br />

Tan, S., Liu, K. and Xie, Z. (2003) A Semiotic Approach to Organisational Modelling Using Norm Analysis,<strong>6th</strong><br />

International <strong>Conference</strong> on Enterprise Information Systems,Porto, Portugal,.<br />

Thanyaphut, A. W., P. (2006) 'Key performance indicators: Strategic into action', Journal of Quality Guarantee, 7(1),<br />

pp.1-12.<br />

Von Krogh, G. and Roos, J. (1996) Managing knowledge: perspectives on cooperation and<br />

competition,Representationism: the Traditional Approach to Cooperative Strategies, London:Sage Publications<br />

Ltd.<br />

Wathne, K., Roos, J. and von Krogh, G. (1996) 'Towards a theory of knowledge transfer in a cooperative context',<br />

Managing Knowledge-Perspectives on cooperation and competition, pp.55-81.<br />

Watson, S. and Hewett, K. (2006) 'A Multi-Theoretical Model of Knowledge Transfer in Organizations: Determinants<br />

of Knowledge Contribution and Knowledge Reuse*', Journal of Management Studies, 43(2), pp.141-173.<br />

Zhihong, L. (2008) A Study on the Factors Influencing Knowledge Transfer among Organizations in Strategic<br />

Alliances,pp.822-825.<br />

Zhou, N., Chen, X.-y. and wang, J.-h. (2006) 'A Study on the design method of KPI system in Chinese<br />

communication enterprises', Journal of Communication and Computer, 6(12), pp.63-69.<br />

111


Exploring the Emotional Exhaustion of Healthcare Providers<br />

Using an In-Hospital Employee Communication Network<br />

Cheng-Yi Chiang 1 , Ya-Ying Kuo 1 , Ying-Hui Hou 2 , and I-Chiu Chang 1<br />

1<br />

Institute of Healthcare Information Management, National Chung Cheng<br />

University, Chiayi, Taiwan<br />

2<br />

Kai Nan University, Taoyuan, Taiwan<br />

g98556002@mis.ccu.edu.tw<br />

yaying.kuo@gmail.com<br />

yhhou@mail.knu.edu.tw<br />

misicc.pig@gmail.com<br />

Abstract: Competition between hospitals in Taiwan is often fierce. Information transparency has replaced the former<br />

information asymmetry situation. Healthcare service providers will have to attract patients in new ways. Healthcare<br />

service providers need to have a positive attitude. The problems of emotional labor and emotional exhaustion of the<br />

front line healthcare givers cannot be over emphasized. It’s important for them to have an emotional support system.<br />

Lifestyles have changed with the development of the Internet and Information Technology. Establishment of an<br />

employee communication network (ECN) in hospitals is essential. Due to the increasing emotional exhaustion of<br />

healthcare providers, this study focuses on one hospital with in-hospital ECN to investigate the extent of healthcare<br />

givers’ emotional exhaustion. This study will further discuss the usage and perceived usefulness of healthcare givers<br />

who have used the ECN. We also hope to ease healthcare givers’ occupational burnout.<br />

Keywords: healthcare providers, nurse, in-hospital employee communication network, emotional exhaustion,<br />

perceived usefulness<br />

1. Introduction<br />

In the 43 major events of Social Readjustment Rating Scale (SRRS) edited by Homels and Ruhe in 1967,<br />

three major events within the top five events are hospital-related ones. The first one is “Death of a<br />

Spouse” with a score of 86. The second one is “Death of Close Family Member” with a score of 77. The<br />

fifth one is “Personal Injury or Illness” with a score of 61. The three events are sources of daily life stress<br />

that will not only influence individuals and families, but also trigger more demands and direct emotional<br />

reactions. Front line healthcare givers who need to face patients or patients’ family members usually bear<br />

considerable blame or negative emotional burdens.<br />

Bolton (2000) mentioned that nurses' 'emotional labor” is hard and productive work and should be valued<br />

in the same way as physical or technical labor. The foundation for organization to benefit comes from<br />

healthcare givers’ emotional labor supply to customers. Organizations are trying to modify and control<br />

staff to show emotions that deliver the expectation of the organization to patients in order to improve<br />

patient satisfaction (Hochschild, 1983).<br />

While healthcare providers face high frequency IT contact and upsurging emotional labors to deal with<br />

daily works, they can employ the virtual in-hospital ECN to: (1) communicate or vent their emotions<br />

through ECN to express negative feelings at the moment and transform emotions. (2) Such<br />

communication methods can temporarily release healthcare givers’ highly emotional labors, so that<br />

negative emotions and effects of healthcare givers can be solved. The stress of healthcare givers can be<br />

reduced to avoid emotional exhaustion. The two points mentioned above are critical keys of this study. If<br />

these two points can be achieved, this can be viewed as a tremendous contribution. For medical<br />

institution executives, they should consider the necessity of building ECN to eliminate job-quitting<br />

problems caused by emotional exhaustion.<br />

2. Literature review<br />

2.1 Emotional labor<br />

Hochschild (1983) first defined “emotional labor” as “the management of feeling to create a publicly<br />

observable facial and bodily display; emotional labor is sold for a wage and therefore has exchange<br />

value.”<br />

112


Cheng-Yi Chiang et al.<br />

Emotional labor must correspond to the following three characteristics: First, workers must create a<br />

publicly observable face-to-face interaction. Second, workers must have an emotional status to effect<br />

others. Third, emotional labors shown by workers are restrained by employees (Wharton, 1993). For<br />

service workers, emotional work has become part of the service (Wichoroski, 1994). Emotional work can<br />

be mostly seen in front-line workers (Ashforth and Humphrey, 1993).<br />

Hochschild (1983) categorized American occupations into high emotion labor occupation and low<br />

emotional labor. Nurses are regarded as professional, technical workers.<br />

2.2 Emotional exhaustion<br />

The definition of Emotional Exhaustion is: a certain reaction which is pressure-related. It is an<br />

overextended emotional phenomenon while interacting with customers or clients. Emotional exhaustion is<br />

often considered as a key element of job stress (Saxton, Phillips and Blakeney, 1991). Jackson, Schwab<br />

and Schuler (1986) believed that workers are involved too much in their jobs. They cited Maslach’s study<br />

and pointed that those work roles involved too much on jobs would have a higher degree in emotional<br />

exhaustion. Such as: workers need to have face-to-face interaction with customers, high emotional<br />

intensity workers or long-term interaction with customers will have a higher degree in emotional<br />

exhaustion. Morris and Feldman (1996) proposed several issues: “frequency of emotional display has<br />

positive correlation on emotional exhaustion”, “attentiveness to emotional display rules has positive<br />

correlation on emotional exhaustion” and “variety of emotions has positive correlation on emotional<br />

exhaustion.”<br />

According to Wharton’s (1993) study focused on bank tellers and hospital workers, it was indicated that<br />

for those who engaged in emotional works have potentially negative consequences of emotional<br />

exhaustion. It depends on authority of job, job involvement and self-control. For high emotional labor<br />

workers with higher authority and self-control are less likely to have emotional exhaustion. Workers with<br />

higher working hours and seniority are likely to have emotional exhaustion. For low emotional labor<br />

workers with higher job involvement, they are less likely to have emotional exhaustion.<br />

The influence of emotional exhaustion is tremendous. Many scholars hope to find out the extraneous<br />

variables of emotional exhaustion to provide a better suggestion and to ease emotional exhaustion<br />

occurrence. Research shows that social supervisory support moderated emotional exhaustion and had a<br />

negative correlation to emotional exhaustion (Dignam et al, 1986; Kirmeyer and Dougherty, 1988).<br />

2.3 Technology Acceptance Model<br />

Technology Acceptance Model (TAM) is designed to understand new information system acceptance<br />

behaviors of IT users. Davis et al. (1989) utilized the “Theory of Reasoned Action” (TRA) as a foundation<br />

and applied information system usage to propose the “TAM.” TAM is mainly to explore the relationship of<br />

the “Perceived Usefulness”, “Perceived Ease of Use”, “Intention to Use” and “Actual System Use” of<br />

information users’ opinions toward certain information systems. TAM claims that “Actual System Use” is<br />

directly determined by “Behavior Intention to Use”; “Behavior Intention to Use” is determined by “Attitude<br />

toward Using” and “Perceived Usefulness”. External variables include: system usage, training, user<br />

intervention of system design process, nature of system design. These variables will effect users’<br />

behavior intentions to use and actual system use indirectly. TRA and TAM both consider that belief will<br />

influence the attitudes of individuals; attitude will influence their willingness and further influence the<br />

behavior performance of individuals.<br />

2.3.1 Hypotheses of TAM<br />

Hypotheses of TAM are as follows (Davis, 1989; Agarwal and Prasad, 1999):<br />

1. Technology usage behavior of humans can be predicted through user willingness of technology<br />

usage. Individual behavior intentions on new technology will be influenced by attitude toward using<br />

new technology.<br />

2. Two determinants of influencing the attitude toward using new technology are “Perceived<br />

Usefulness” and “Perceived Ease of Use.” “Perceived Usefulness” is the key factor for individuals to<br />

accept new technology. “Perceived Ease of Use” is the secondary influencing factor for individuals to<br />

accept new technology.<br />

113


Cheng-Yi Chiang et al.<br />

3. “Perceived Usefulness” influences not only “Attitude toward Using” but also “Intention to Use.”<br />

4. “Perceived Ease of Use” will effect individuals’ perception on “Perceived Usefulness.” The higher<br />

degree of “Perceived Ease of Use” is, the higher individuals’ perception on “Perceived Usefulness”<br />

toward new technology will be.<br />

5.”External Variables” means external factors that will potentially influence users “Perceived<br />

Usefulness” and “Perceived Ease of Use.” Such as individual differences, system character and<br />

environments. These external variables will influence user perceptions on “Perceived Usefulness”<br />

and “Perceived Ease of Use.” (Venkatesh and Davis, 1996; Szajna, 1996).<br />

2.3.2 Revised Technology Acceptance Model<br />

Not many concrete studies were done in understanding the relationship among “Perceived Usefulness”,<br />

“Attitude toward Using”, “Intention to Use”, and “System use”, which resulted in debates of many<br />

scholars. Scholars like Adams et al, (1992) and Straub et al, (1996) have proposed the revised model of<br />

TAM. Their findings suggest that the “Behavior Intention to Use” of TAM should be removed. Igbaria<br />

(1997) further applied the revised TAM on the issue of technology acceptance in small enterprises. The<br />

IT discussed in this study (in-hospital ECN) is a website which is built by the hospital itself. It is not<br />

complicated as other information systems and healthcare givers are capable of operating Internet.<br />

Hence, this study only takes “Perceived Usefulness” as its research variable.<br />

3. Research method<br />

3.1 Objectives and sampling method<br />

The research method of this study is a multi-case study. Questionnaires were dispatched to one juridical<br />

person medical center and healthcare givers of regional hospitals. Healthcare givers include all-levels of<br />

staffs and executives. The sampling method of this study is a stratified random sampling. Distribution of<br />

questionnaires is calculated by the ratio of healthcare givers in the three medical institutions. 80, 850, and<br />

70 questionnaires were given to the three medical institutions, respectively. Method of questionnaire<br />

distribution: A workshop for this questionnaire will be held in one week before the collection date.<br />

Questionnaires will be given to all divisions by this study’s researcher to ask for lower nursing managers’<br />

assistance in the questionnaire distribution. When questionnaires are completed, they will be sent back<br />

by lower nursing managers or be collected by researcher of this study.<br />

3.2 Research framework and hypotheses<br />

This study hopes to find out the relativity of emotional exhaustion through perceived usefulness, usage<br />

pattern and motivation of using the in-hospital ECN. The structure of this study is shown as Figure 1.<br />

H1 H2<br />

Figure 1: Structure of this study<br />

H1: In-hospital healthcare givers’ perceived usefulness toward ECN has significant correlation with using<br />

experience.<br />

H2: The experience of using this ECN has significant correlation with emotional exhaustion.<br />

3.3 Questionnaire design<br />

This questionnaire in this study is a structure questionnaire. Variables include: the degree of emotional<br />

exhaustion of healthcare givers, personal information, experiences of using in-hospital ECN, perceived<br />

114


Cheng-Yi Chiang et al.<br />

usefulness of using this website, perceived usefulness, character of in-hospital ECN and one<br />

open-ended question which is suggestions to this website. The extent of emotional exhaustion and<br />

perceived usefulness are continuous variables. Points are measured by the use of the Likert 7-point<br />

scale.<br />

3.3.1 Validity analysis<br />

A structured questionnaire is being used in this study. Some questions in Maslach and Johnson (1981)<br />

and Davis’ studies (1989) have been used in this questionnaire. Seven general experts with experience<br />

in the field (3 of them are Information Management experts and the others are psychiatric experts) will be<br />

invited to check and review the 21 continuous variables for its adequateness and necessity. A five-point<br />

scale is being used as its rating scale. Questions higher than 4 points will be adopted and validity<br />

measurement of experts have an average score of 4.5 points.<br />

3.3.2 Reliability analysis<br />

This study employs Cronbach’s α value to measure the consistency of every question,Non-research<br />

samples will be chosen for testing when questionnaires are examined for content validity. Modification will<br />

be done based on respondents’ opinions when questionnaires are being collected. The value of<br />

Cronbach’s α is more than 0.8 which stands for a high validity. The whole scale was 0.917 for emotional<br />

exhaustion and the Cronbach’s α of perceived usefulness of ECN was 0.971. Both of them reached<br />

significant validity.<br />

3.4 Data analysis method<br />

3.4.1 Descriptive statistics<br />

Descriptive statistics is used to analyze the perceived usefulness, using experience, motivation,<br />

emotional exhaustion awareness, and characteristics of the ECN. Normal distribution and usage<br />

distribution will also be included in the statistics process and multiple choice questions will be utilized to<br />

analyze the motivation and characteristics of in-hospital ECN.<br />

3.4.2 One-Way MANOVA<br />

One-Way MANOVA is mainly used to explore the correlation of the using experiences and emotional<br />

exhaustion of in-hospital ECN and the correlation of variables of using experience and perceived<br />

usefulness.<br />

3.4.3 Independent t-test<br />

Independent t-test is used to analyze the correlation of emotional exhaustion and perceived usefulness of<br />

in-hospital ECN.<br />

3.4.4 Pearson correlation<br />

Pearson correlation is employed to examine the correlation of healthcare givers’ experiences of<br />

in-hospital ECN with the perceived usefulness and emotional exhaustion awareness.<br />

4. Results and discussion<br />

1000 questionnaires were dispatched and were numbered from 1 to 1000. Questionnaires will be<br />

collected two weeks after distribution. Of the 1000 dispatched questionnaires, 918 were returned for a<br />

91.8% return rate. Effective respondent rate is 91.1%.<br />

4.1 Descriptive statistics<br />

4.1.1 Using experience<br />

The analysis of using experience: (1) Continuous users take the majority for 73.47%. 3/4 of the<br />

healthcare givers still keep using this website. (2) Users with five years or above using experiences have<br />

rate of 35.6% and users with 2-3 years of using experience have a rate of 13.7%. (3) Frequency of using<br />

ECN is mostly about 1 to 3 times. (4) Time of use is predominantly within 5 minutes (63.9%) and if those<br />

using time within 5 to 15 minutes are included in the calculation; the rate of time of use within 5 to 15<br />

115


Cheng-Yi Chiang et al.<br />

minutes will be 93.6%. (5) The main reason for never using this website is lack of interest (57%) and<br />

other reasons of never using this website has a rate of 21.9%.Non-continuous users have a rate of 64.5%<br />

and the main reason is lack of time. Void of content comes to the next which has a rate of 14.0%. Other<br />

reasons of using this website are ordered from: “lack of time”, “worried of being interviewed by<br />

supervisors and privacy of information exchanging is not secured.”<br />

4.1.2 Motivation<br />

Multiple choice percentages of motivation analysis is as follows: (1) Entertainment (19.6%) (2)<br />

Information sharing (18.5%) (3) Emotional relief (17.2%) (4) Learning (15.7%) (5) Emotional support<br />

(13.0%). The main reason for using this website us for entertainment and reason related to emotional<br />

relief and emotional support has a rate of 30.2%. 10.9 of the respondents use this website to monitor the<br />

condition of hospital.<br />

4.1.3 Characteristics of In-hospital ECN<br />

Characteristics of In-hospital ECN are ordered from: (1) Variety of topics (31.2%) (2) Enthusiastic<br />

interaction among members (23.6%) (3) Privacy of information (18.1%) Almost one-third of the<br />

respondents will like to have various topics for information exchanging and one-seventh (15.1) of the<br />

respondents will like to receive feedback for hospitals.<br />

”Other factors” in Analysis Chart of Characteristics of In-hospital ECN has a rate of 2.4%. Most opinions<br />

in the open-ended question are “hoping questions raised can be responded positively by hospitals”,<br />

“worried of being interviewed by supervisors, lack of privacy, hoping information could be secured.”<br />

4.2 Perceived usefulness and emotional exhaustion distribution<br />

4.2.1 Perceived usefulness of Healthcare Givers toward In-hospital ECN<br />

Perceived Usefulness and Emotional Exhaustion Distribution of Healthcare Givers toward In-hospital<br />

ECN are ordered from: (1) Beneficial to solve daily problems and with a score of 4.19. (2) Useful in daily<br />

jobs and with a score of 4.18. (3) Ease patients’ unpleasant emotion. (4) Less stressed in work emotion<br />

and with a score of 4.11. (5) Expressing ideas or suggestions to hospitals and with a score of 4.09.<br />

Average perceived usefulness score toward in-hospital ECN is 4.0 and only 25% of the respondents have<br />

an average score less than 4.0. Mode of the respondents is 5.0.<br />

4.2.2 Emotional exhaustion of Healthcare Givers<br />

The distribution of respondents’ emotional exhaustion is ordered from: (1) “Exhausted after duty” has the<br />

highest average score of 5.69. (2) “Emotional exhaustion of daily work” comes the next and with an<br />

average score of 5.20. (3) “Seems to have occupational burnout” ranked the third and with an average<br />

score of 5.02. (4) “Feel stressed when contact with customers” ranked the lowest score of 3.99.<br />

The average score of respondents’ emotional exhaustion is 4.80 and only 25% of the respondents have<br />

an average score less than 4.0. Mode of the respondents is 5.0. This phenomenon reveals that the e<br />

degree of emotional exhaustion of healthcare givers is relatively high.<br />

4.3 Related analysis<br />

4.3.1 Correlation between whether healthcare givers use the In-hospital ECN and emotional exhaustion<br />

awareness<br />

Whether healthcare givers use the in-hospital ECN or not has significant correlation with emotional<br />

exhaustion awareness. It conceals that there will be a significant difference in emotional exhaustion when<br />

healthcare givers do not use the website. Healthcare givers who employ the in-hospital ECN have lower<br />

emotional exhaustion awareness. Hence, H2 above of this study is supported; see Table 1.<br />

116


Cheng-Yi Chiang et al.<br />

Table 1: Independent sample t-test analysis of whether healthcare givers use in-hospital ECN and<br />

emotional exhaustion<br />

Item Use ECN Number Mean SD Standard error of mean<br />

Sum<br />

No 189 45.132 10.4402 .7594<br />

emotional<br />

exhaustion Yes 614 42.800 10.9403 .4415<br />

Levene equal<br />

variance test<br />

T test of equal mean<br />

F test Significant t df Significant Standard<br />

error<br />

Sum emotional exhaustion Assuming<br />

equal<br />

variance<br />

2.123 .146 2.590 801 .010 .9005<br />

Not<br />

assuming<br />

equal<br />

variance<br />

2.655 325.178 .008 .8784<br />

4.3.2 Correlation between Healthcare givers’ perceived usefulness and using experience toward<br />

in-hospital ECN<br />

This study is shown as Table 2:<br />

1. The perceived usefulness has a significant correlation with frequency of using the website. This<br />

means that the perceived usefulness of In-hospital ECN for healthcare givers is different in statistics,<br />

so is its frequency. Hence, the above H1 of this study is supported.<br />

2. The perceived usefulness has significant correlation with time of use. This shows that when the<br />

perceived usefulness is different in statistics and there will be great difference on time of use. The<br />

above H1 is supported.<br />

Table 2: Perceived usefulness and using experience and Pearson correlation analysis chart<br />

Usefulness<br />

Use frequency<br />

Use time<br />

5. Conclusions and suggestions<br />

Usefulness Use frequency Use time<br />

Pearson correlation 1 .331(**) .388(**)<br />

Significant .000 .000<br />

Number 804 804 801<br />

Pearson correlation .331(**) 1 .663(**)<br />

Significant .000 .000<br />

Number 804 835 832<br />

Pearson correlation .388(**) .663(**) 1<br />

Significant .000 .000<br />

Number 801 832 832<br />

1. The degree of emotional exhaustion and the mode of occupational burnout are greater than<br />

average value. It appears that healthcare givers are stressful at work and supervisors should take<br />

such problems seriously to avoid follow up questions.<br />

2. Reasons for those who do not use this website are: “lack of time “is the main reason; “worried of<br />

being interviewed and information is not secured “ranked the second reason. Healthcare givers are<br />

lack of time to use in-hospital ECN instead of lack of willingness. Some consider that the website is<br />

void of content and time-wasting. Moreover, it can be verified by the value of frequency and time of<br />

use. (Frequency and time of use are of low value)<br />

3. The main motivation of using this website is for entertainment; however, 30% of the respondents<br />

use this website to seek for emotion expression and emotional support. It is clear that healthcare<br />

givers still need a way to vent their feelings. 30% of the respondents use this website for<br />

information sharing and learning. Healthcare givers have the motivation of learning new knowledge. If<br />

this virtual space is being utilized properly by supervisors, healthcare givers can have their<br />

information exchanged and reached the goal of learning new knowledge and information sharing. For<br />

117


Cheng-Yi Chiang et al.<br />

the motivation of “monitoring the whole hospital”, healthcare givers expect to realize the policy and<br />

decision of hospitals and obtain related answers.<br />

4. Average score of perceived usefulness is 4.0. “Daily work assistance” scores higher among all the<br />

items within the questionnaire of in-hospital ECN. It reveals that healthcare givers can solve daily<br />

problems by the use of information sharing and correspond to the “information sharing” of motivation.<br />

Perceived usefulness of “less stressed at work”, ”releasing unpleasant emotion while taking care of<br />

patients” and “less frustration at work ”corresponds with “emotional expression and emotional<br />

support” of motivation in this study. Perceived usefulness of “feel stressed while interacting with<br />

patients”, “emotional exhaustion” and “exhausted” can still not be improved by the use of in-hospital<br />

ECN. Perceived usefulness of” same as solving working problems” can be traced back to the<br />

biggest demands of healthcare givers. Unpleasant emotion or energy of healthcare givers need to be<br />

released timely. It will be effective even if healthcare givers merely use this website to express their<br />

feeling after work. Hence, in-hospital ECN is relatively critical.<br />

5. Healthcare givers are expecting the in-hospital ECN to have characteristics as follows: “varied<br />

topics”(31.2%), “enthusiastic interaction”(23.6%), “information to be secured”(18.1%), “feedback from<br />

hospitals”(15.1%) and “stability of this website”(9.7%) The great concern of healthcare givers is the<br />

diversification of in-hospital ECN; not only merely emotion expression. This finding conforms to<br />

references and this study. Healthcare givers hope to build a multi-functional virtual space and this<br />

opinion has been mentioned in the open-ended question. In the aspect of “feedback from hospitals”,<br />

questions raised by healthcare givers did not receive adequate responses; it ended up with opinions<br />

of “worried of being interviewed by supervisors”, ”information is not being secured” and “lack of<br />

privacy.” Supervisory agencies only scold healthcare givers instead of understanding their thoughts<br />

and feelings. If questions raised by healthcare givers could be bulletined in public, healthcare givers<br />

and hospital executives can understand the other party’s needs and expectations. Hence, the<br />

establishment of this website will be meaningful. Executive should be more tolerating to avoid<br />

positive suggestions being neglected. Hospital should create a virtual space for healthcare givers to<br />

express their thoughts and ideas. In other words, healthcare givers will no longer use the in-hospital<br />

ECN to express their genuine feelings.<br />

The average score of emotional exhaustion is 5.0. However, objectives of this study are first line<br />

healthcare givers and executives. Their sources of pressure are different than those of high-level<br />

executives. The degrees of emotional exhaustion may also differ in the source of pressure. Emotional<br />

exhaustion of high-level executives should be explored further to serve as a reference of emotional<br />

expression.<br />

This study only discusses the using experience and perceived usefulness of in-hospital ECN. Content of<br />

information exchanging is not being explored in this study. Follow-up researchers can probe further the<br />

content of information exchanging of in-hospital ECN for further application to practical field and<br />

academia. (Ex: turnover prediction, work-related stress, work satisfaction, burnout prediction and<br />

e-counseling service etc.)<br />

References<br />

Adams, D.A., Nelson, R.R. and Todd P.A. (1992) “Perceived Usefulness, Ease of Use, and Usage of Information<br />

Technology:A Replication”, MIS Quarterly, Vol. 16, No.2, pp 227-247.<br />

Agarwal, R. and Prasad, J. (1999) “Are Individual Differences Germane to the Acceptance of Information<br />

Technology. Decision Sciences, Vol. 30, No. 2, pp 361-391.<br />

Ashforth, B.E. and Humphrey, R.H. (1993) “Emotional Labor in Service Roles:The Influence of Identity”, <strong>Academic</strong> of<br />

Management Review, Vol. 18, No. 1, pp 88-115.<br />

Bolton, S. (2000) “Who Cares Offering Emotional Work as a ‘gift’ in the Nursing Labor Process”, Journal of Advanced<br />

Nursing, Vol. 32, No.3, 580-586.<br />

Davis, F. (1989) “Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology”,<br />

MIS Quarterly, Vol. 13, No. 3, pp 319-340.<br />

Davis, F.D., Bagozzi, R.P. and Warshaw, P.R. (1989) “User Acceptance of Computer Technology:A Comparison of<br />

Two Theoretical Models”, Management Science, Vol. 35, No. 8, pp 982-1003.<br />

Dignam, J.T., Brarrera, M.Jr. and West, S.G. (1986) “Occupational Stress, Social Support, and Burnout among<br />

Correctional Officers”, American Journal of Community Psychology, Vol. 14, No. 2, pp 177-193.<br />

Hochschild, A. R. (1983) The Managed Heart: Commercialization of Human Feeling. Berkeley: University of<br />

California Press.<br />

Holmes, T. and Rahe, R. H. (1967) “The Social Readjustment Rating Scale”, Journal of Psychosomatic Research,<br />

Vol. 11, No. 2, pp 213-218.<br />

118


Cheng-Yi Chiang et al.<br />

Igbaria, M., Zinatelli, N., Cragg, P. and Cavaye, A.L.M. (1997) “Personal Computing Acceptance Factors in Small<br />

Firms: A Structural Equation Model”, MIS Quarterly, Vol. 21, No. 3, pp 279-305.<br />

Jackson, S.E., Schwab, R.L. and Schuler, R.S. (1986) “Toward an Understanding of the Burnout Phenomenon”,<br />

Journal of Applied Psychology, Vol. 71, No. 4, pp 634-640.<br />

Kirmeyer, S. L. and Dougherty, T. W. (1988) “Work Load, Tension, and Copying: Moderating Effects of Supervisor<br />

Support”, Personnel Psychology, Vol. 41, No. 1, pp 125-139.<br />

Maslach, C. and Jackson, S. E. (1981) “The Measurement of Experienced Burnout”, Journal of Organizational<br />

Behavior, Vol. 2, No. 2, pp 99-113.<br />

Morris. J. A. and Feldman, D. C. (1996) “The Dimensions, Antecedents, and Consequences of Emotional Labor”,<br />

The Academy of Management Review, Vol. 21, No. 4, pp 986-1010.<br />

Saxton, M. J., Phillips J. S. and Blakeney, R. N. (1991) “Antecedents and Consequences of Emotional Exhaustion in<br />

the Airline Reservations Service Sector”, Human Relations, Vol. 44, No. 6, pp 583-595.<br />

Straub, D., Limayem, M. and Karahanna-Evaristo, E. (1995) “Measuring System Usage : Implications for IS Theory<br />

Testing”, Management Science, Vol. 41, No. 8, pp 1328-1342.<br />

Szajna, B. (1996) “Empirical Evaluation of the Revised Technology Acceptance Model”, Management Science, Vol.<br />

42, No. 1, pp 85-92.<br />

Venkatesh, V. and Davis F. (1996) “A Critical Assessment of Potential Measurement Biases in the Technology<br />

Acceptance Model: Three Experiments”, International Journal of Human-Computer Studies, Vol. 45, No. 1, pp<br />

19-45.<br />

Wharton, A.S. and Erickson, R.J. (1993) “Management Emotions on the Job and at Home:Understanding the<br />

Consequences of Multiple Emotional Roles”, The <strong>Academic</strong> of Management Review, Vol. 18, No. 3, pp<br />

457-486.<br />

Wichoroski, M.A. (1994) “The Secretary: Invisible Labor in the Workworld of Women. Human Organization, Vol. 53,<br />

No. 1, pp 33-41.<br />

119


The Implementation of RSS-Based Clinical Reminder<br />

Wen-Chou Chi 1 , Chia-Hsien Wen 2 , Sek-Kwong Poon 3 and Shih-Che Huang 3<br />

1 National Chung Cheng University, Chiayi, Taiwan<br />

2 Providence University, Taichung, Taiwan<br />

3 Taichung Veterans General Hospital, Taichung, Taiwan<br />

y6312002@gmail.com<br />

chwen@pu.edu.tw<br />

Abstract: Computer base reminder system can help physicians to get right information and make right decision in<br />

daily clinical work. This study presents a RSS-based Clinical Reminding System (RCRS) that was designed and<br />

implemented for reminding clinicians to deal with their varied unfinished clinical works. RCRS provides XML-formatted<br />

RSS clinical messages through automatically collecting clinical information for every clinician from the hospital<br />

information system (HIS). There are two types of system structure in our design, one is stand along module which is<br />

working in specific PC, and the other is web-based module, integrated with the CPOE system. According to this<br />

design, clinicians will receive and read these messages whenever he/she logs on the HIS. A hyperlink, which<br />

connects to appropriate applications and let the clinician easily make some response, was provided by each RSS<br />

message. The system can help clinicians focus on patient care without keeping track of clinical chores. Information<br />

“Content” of this system was suggested to modify by system reviewers, but information “Accuracy”, “Formats”, “Ease<br />

of use” and “Timeless” of this system is appropriate for system design purposes.<br />

Keywords: RSS, CPR, XML, clinical reminder<br />

1. Introduction<br />

Information system has been applied in medical service for a while, there are two important contributions<br />

of these computer-based service systems: reduce medical error and increase service efficiency (Aspden,<br />

et al., 2006). Because the huge database conduct by imputed clinical orders and Electronic Medical<br />

Record (EMR), there are a lot useful information in the system and these information will help us to do<br />

appropriate decision in clinical service through Clinical Decision Support System (CDSS). According the<br />

request proposed by clinical staff, CDSS use the data base form EMR and the algorithms that made by<br />

evident to provide some recommendation to these staff.<br />

The computer reminder is one of the CDSS used in hospital whose role is more and more important in<br />

medical service. Physicians can get important information from the computer reminder and use these<br />

information to do decision. But most reminder system are individual system, it can only provide one single<br />

event alert (Hunt, Haynes, Hanna and Smith, 1998). Recently, some reminder systems start to use in<br />

disease management or prevent disease, which means this system start to provide “a series of alert”, but<br />

still a single system (Garg, et al., 2005). In the clinical daily work, the physicians need a lot of information<br />

from many subsystem of Hospital Information System (HIS), every subsystem have its own alert, if we can<br />

collect all these information, not just one event or one disease, and provide physicians an “to do list”, it<br />

would be helpful in prevent error and increase service quality.<br />

The effeteness of computer reminders is debatable, some studies conclude that computer reminders<br />

produced much smaller improvements than those generally expected from the implementation of<br />

computerized order entry and electronic medical record systems (Shojania, et al., 2010); the others<br />

conclude in opposite way (Hunt, Haynes, Hanna and Smith, 1998; Garg, et al., 2005). Infect, the useful<br />

reminder has few characteristics (Lobach, 2005): automatically prompted to use the system; integration<br />

with charting or order entry system and response required. RSS, which stands for Really Simple<br />

Syndication, Rich Site Summary or Resource Description Framework (RDF) Site Summary, is a data<br />

exchange protocol for sharing web contents, may fit the requires of clinical reminder system. It may<br />

simplify and speed the users’ access to the latest updated information which is published by a web site.<br />

The publisher web site creates an RSS file, also known as RSS feed or RSS channel, in which the content<br />

to be published is transformed into an XML-based format. Any user can read the RSS feed at different<br />

sites at the same time by way of an RSS aggregator (RSS reader).<br />

An RSS aggregator is an application providing the means to read the content of RSS file before it can be<br />

displayed (Cadenhead, Curry and Zellers, 2005a; Bray, et al., 2004; Ola and Niclas, 2005). In general,<br />

there are three common types of RSS aggregators. The first type RSS aggregator, Desktop RSS<br />

aggregator, is the standalone program. Users download this program to their computers from Internet and<br />

use it to collect the feeds and refresh items when a feed is updated. Web-based aggregator, the second<br />

120


Wen-Chou Chi et al.<br />

type, is a kind of online web service. Users can use it to collect the feeds and browse messages in<br />

Internet. The final type is Plug-in aggregator which allows users to view RSS feeds with an existing<br />

program such as Microsoft Outlook (Ola and Niclas, 2005). RSS is generally used in real-time news<br />

reading and publication information management (Hu, Lim and Rizos, 2005; Klein, Mika and Serban;<br />

Çelikbaş). There are some RSS-based applications used in health care territory, but no one is applied in<br />

hospital based medical care (Phillippi and Buxton, 2010; Ku and Huang, 2010).<br />

The proposed RSS-based clinical information reminder is implemented under the support of a medical<br />

center in Taiwan. The reminder is coupled with the HIS. It will periodically collect clinical information<br />

required by clinicians from HIS and transform them into RSS messages automatically. A clinician will<br />

receive these RSS clinical messages sent to him/her whenever he/she logs on HIS. The reminder is also<br />

associated with the LDAP server of HIS for user authentication to protect patient privacy.<br />

RSS is an effective tool to spread internal information. One of the functional components of a CPR system<br />

is integrated communication support. A patient-oriented medical record will be the communication tool for<br />

team members who take care of a patient. Traditional computer-based patient-record (CPR) systems<br />

generally provide passive pull-type communication because each team member has to scan through the<br />

record to find other’s opinion. By including RSS mechanism in a CPR system, each team member’s<br />

opinion will automatically push to other members.<br />

2. System design<br />

We have built the RSS-based clinical information reminder. In order to help user operation be more<br />

effective and efficient, we designed and built two modes for user operation in our system: one is the<br />

standalone system which installs the PCs in the clinic and station; the other is the web-based module<br />

which integrates with the CPOE system. The system architecture has shown schematically in figure 1,<br />

uses the following components and sources of clinical information.<br />

Figure 1: The system architecture and its operation flow of RSS-based clinical information reminder<br />

2.1 Standalone mode<br />

In standalone mode, PCs and Workstations installed in the standalone system for front-end of reminder<br />

and used by healthcare workers to login in the reminder through the intranet. LDAP Server receives the<br />

user’s login request and identifies the user’s identity and authority. In RSS reminder, the patient’s<br />

information only allows the related healthcare workers to access. HIS which stores the clinical information<br />

and translated information to XML-formatted files. RSS Server includes the RSS aggregator and the<br />

121


Wen-Chou Chi et al.<br />

database server of the reminder. The reminder works by the way of coupling itself with the HIS of a<br />

hospital. The RSS server periodically collects XML-formatted clinical information from HIS and saves them<br />

into the RSS database as RSS clinical messages. The RSS feed is created concurrently.<br />

2.2 Web-based mode<br />

In our case, the web-based mode is similar to the standalone mode, but there are some differences in<br />

them. In this mode, we develop a RSS web service module and integrate with the CPOE system because<br />

the clinicians are familiar with CPOE in the hospital. When clinician uses the RSS service in the CPOE,<br />

the RSS web service will send the clinician id and patient’s id to the RSS server through the delivered web<br />

address. RSS server receives the address and interprets it, then returns the patient’s clinical information<br />

to the RSS web service. The example of web address has shown as<br />

http://xxx.xxx.xxx./?doctor_id=Y123009 &patient_id=A123008772.<br />

2.3 Data structure of the reminder<br />

Figure 2 shows an example of RSS clinical messages stored in the clinical information reminder. The data<br />

structure of the reminder consists of an index-file and many data-files.<br />

Figure 2: An example of clinical messages stored in the clinical information reminder<br />

The index-file is the RSS feed that lists some important attributes, such as patient id, data-file name, and<br />

physician id of all data-files. A data-file records detailed data of a clinical information. Both of the index-file<br />

and data-files are in XML format conformed to RSS 2.0. A element of the RSS feed represents<br />

a class of RSS clinical messages. There will be n elements in a channel, if there are n RSS clinical<br />

messages in the class. The patient id, data-file name, and physician id of a RSS clinical message are<br />

122


Wen-Chou Chi et al.<br />

correspondingly stated in element, element, and element of an item. Since a<br />

physician will not always use a specific client, the RSS aggregator of the reminder is designed as a<br />

web-based, the second type, aggregator. The RSS server performs event-driven RSS syndication,<br />

because only those messages with matched doctor’s id will be retrieved and displayed to the user (Ola<br />

and Niclas, 2005; Coile, 2002). Whenever a physician logs in HIS, the LDAP server will pass the address<br />

of the RSS aggregator and ciphered user data to the RSS server to initiate the reminder after<br />

authentication. The aggregator will get the RSS feed and pick up those items whose doctor’s id in<br />

element match with the one obtained from the LDAP server. A matched item represents a<br />

sound message. Then the aggregator passes the list of sound messages to the RSS server and requests<br />

it to retrieve sound messages from RSS database. All sound messages are displayed for advanced<br />

process by the clinician.<br />

3. System implementation<br />

The reminder is developed using the VB.net and ASP.net programming language and Microsoft SQL<br />

Server 2000. In our system, there are four classes of RSS clinical messages currently processed in the<br />

reminder: Abnormal laboratory results notification, abnormal examination results notification, discharge<br />

note notification, and consultation notice notification. Other kinds of clinical information will be joined to the<br />

reminder gradually.<br />

An abnormal laboratory results notification messages is sent to notify a physician that the result of a<br />

laboratory test he/she ordered is abnormal and critical. If a test result is verified to be abnormal and out of<br />

the acceptable range set in advance, relative information about the abnormal laboratory result, such as<br />

patient’s name, the laboratory result, and the order date-time will be immediately sent to the physician who<br />

ordered this test. Figure 3 shows an example screen about abnormal laboratory results notification in the<br />

standalone mode.<br />

Figure 3: An example screen about abnormal laboratory results notification<br />

The left part of the screen is a class list of RSS clinical messages. Archived messages will be shown<br />

Archived messages will be shown while “archived messages” item in the class list was clicked. The upper<br />

right part of the screen lists laboratory results notification messages that the clinician was received. If one<br />

of the messages in the list is clicked, detail information of the message will be shown on the lower right<br />

part of the screen. The physician may enter into laboratory information system by clicking the asterisk in<br />

the “link” field of the message.<br />

Figure 4 shows an example screen about the abnormal examination results notification. Similar to the<br />

appeal of abnormal laboratory results notification, it is used to notify a physician that the result of an<br />

imaging examination or a pathology examination he/she ordered is abnormal.<br />

123


Wen-Chou Chi et al.<br />

Figure 4: An example of transforming an abnormal examination results notification message into an<br />

archived one<br />

Discharge note notification is used to notify a clinician that one of his/her patient is discharged and he/she<br />

has to accomplish the discharge note before deadline. Figure 5 shows the RSS clinical messages about a<br />

clinician’s discharge note notification.<br />

Figure 5: An example of deleting a discharge note notification message<br />

A consultation notice notification message notifies a clinician that there is a consultation he/she has to<br />

hold. An example screen about consultation notice notification is shown in figure 6.<br />

4. Evaluation<br />

The case hospital is a medical center in central Taiwan. At present, the case hospital has more than 500<br />

physicians and 1,500 beds. The project team members come from the information department and clinical<br />

informatics research & development center (CIRD). Two project leaders are director from the CIRD.<br />

During the reminder design stage, they were responsible for discussing the reminder functions in the<br />

physicians group meeting. After reminder was implemented, they were responsible to train and collect<br />

feedback from the physicians. So, we chose the two project leaders as our interviewee to give a<br />

124


Wen-Chou Chi et al.<br />

preliminary evaluation of the system. The interview results were recorded using the taxonomy developed<br />

by Doll and TorKzadeh (1998) in The Measure of End-User Computing Satisfaction.<br />

Figure 6: An example screen about consultation notice notification<br />

The interviewees summarized that most users suggested that the “content” of the system needs to be<br />

screened in order to avoid information overload. Because the information provided by the system is not<br />

“created” from the system, but “collected” from other system, both evaluators believed the information<br />

provided by the system is “Accurate”. Regarding to the “Ease of use” and “Timeless” of the system, the<br />

evaluators scored “satisfied” with two recommendations. Firstly, they suggested that the end users need<br />

more authorizations to manipulate the information they want, because some events may be meaningful for<br />

some users but not others. Secondly, use the remote technology to deliver message to end user may<br />

increase the accessibility of the system.<br />

5. Discussion<br />

Providing too many information will cause the physician “information overlord” (Rebitzer, Rege and<br />

Shepard, 2008), so do this system. In the future, if this system can screen information and verify which<br />

information are important and which are not, it would be helpful. Unfortunately, it need time to build up this<br />

screen system.<br />

Because of the information which physicians need were come from different sub system in hospital<br />

information system, all the reminding system need is collect information from different subsystem. RSS<br />

has these features, the user don’t have to pick up information around hospital, the information we need will<br />

active “feed ” the RSS reminding system. There are two advantage of this system, more accuracy of these<br />

information and more easy to get information.<br />

The amount of control for the system is considered difficult to determine. Too much control by users<br />

makes it difficult to follow, and too little control limits the control by users (Sprague, 1980). In order to<br />

prevent the loss of essential information, RSS does not allow users to determine the type of information<br />

provided by the program. Multi-disciplinary care is usually involved in patient care. Before the consensus<br />

was made, RSS provided standardized information for all professionals involved in the patient care.<br />

In order to reduce the amount of errors in patient care, getting correct information on time is important.<br />

Although RSS is able to provide information on time, the medical professionals need to be using the<br />

specific computers to access the information. The medical professionals rarely stay in front of a computer<br />

during work. Therefore, it is important to develop a system that is capable of sending real-time information<br />

to the medical professionals via wireless devices such as mobile phones.<br />

125


6. Conclusion<br />

Wen-Chou Chi et al.<br />

RSS provides a mean to acquire up-to-date information quickly from Internet. According to the evaluation<br />

by project leaders, this RSS-base reminder system is useful and effective and suggested to the<br />

superintendent not only to expand the RCRS to include all kinds of messages needed by the physicians<br />

but also put this idea into the design of the next generation Hospital Information System. They also point<br />

out that physicians may be overburdened if provided too much information may not be relevant. If this<br />

system can screen information and verify important information would be helpful and increase the users’<br />

acceptance of this system. This paper presents the design and implementation of an RSS-based clinical<br />

information reminder. A clinician may use the reminder to receive all kinds of clinical messages and<br />

arrange them as a to-do list. Although the pioneer system only includes four classes of clinical messages,<br />

it proves that the reminder will help clinicians in managing clinical chores.<br />

Acknowledgments<br />

This paper is based upon work supported by National Science Council (NSC), Taiwan under grants<br />

no. NSC96-2221-E-126-006, NSC96-2221-E-126-004-MY3 and NSC95-2218-E-007-025. Any opinions,<br />

findings, and conclusions or recommendations expressed in this material are those of the authors and do<br />

not necessarily reflect the views of the NSC.<br />

References<br />

Aspden, P. et al., (2006) Committee on Identifying and Preventing Medication Errors. Preventing medication errors:<br />

quality chasm series. Washington(DC): The National Academies Press.<br />

Bray, T. et al., Extensible Markup Language (XML) 1.0 (Third Edition) W3C Recommendation 04 February 2004, The<br />

World Wide Web Consortium, 2004, last visited 2005-02-18, http://www.w3.org/TR/2004/REC-xml-20040204/<br />

Cadenhead, R. Curry, A. and Zellers, S. (2005a) “RSS at Harward Law”,[online], RSS Advisory<br />

Board, http://blogs.law.harvard.edu/tech/rss#whatIsRss<br />

Çelikbaş, Z. “What is RSS and how can it serve libraries?” Learning for the Future: e-Learning, pp. 277-292, İstanbul<br />

(Turkey).<br />

Coile, R.C. Jr. (2002) “The Paperless Hospital, Healthcare in a Digital Age,” Foundation of the American College of<br />

the Healthcare Executives.<br />

Doll, W.J., TorKzadeh G.(1998) "The Measure of End-User Computing Satisfaction. Management Information<br />

Systems Quarterly." Vol. 12, No.2, pp 259-274.<br />

Garg, A.X. et al., (2005) "Effects of computerized clinical decision support systems on practitioner performance and<br />

patient outcomes: a systematic review." JAMA, Vol 293, No. 10, 9 Mar, pp 1223-1238.<br />

Hu, Y.H. Lim, S. and Rizos, C. (2005) “Delivering GNSS Data Over the Internet Using RSS for Post-processing<br />

Applications” International Symposium on GPS/GNSS.<br />

Hunt, D.L. Haynes, R.B. Hanna, S.E. Smith, K. (1998)” Effects of computer-based clinical decision support systems<br />

on physician performance and patient outcomes: a systematic review.” JAMA, Vol 280, No. 15, 21 Oct, pp<br />

1339-1346.<br />

Klein, M. Mika, P. and Serban, R. “Semantics-based Publication Management using RSS and FOAF”<br />

SemanticDesktop.org.<br />

Ku, H.H. and Huang, C.M. (2010) "Web2OHS: A Web2.0-Based Omnibearing Homecare System." IEEE transactions<br />

on information technology in biomedicine, VOL. 14, NO. 2,<br />

Lobach, D.F. (2005) "Improving clinical practice using clinical decision support systems: a systematic review of trials<br />

to identify features critical to success." BMJ, Vol 330, No.7494, 2 Apr, pp 765.<br />

Ola, A. and Niclas, (2005) L. “RSS – The future of internal communication?” Reports from MSI, ISSN 1650-2647.<br />

Phillippi, J. C. and Buxton, M. (2010) "Web 2.0: Easy Tools for Busy Clinicians." Journal of Midwifery & Women’s<br />

Health. Vol 55, No. 5, pp 472-476.<br />

Rebitzer, J. B., Rege, M., and Shepard, C.(2008) "Influence, information overload, and information technology in<br />

health care." Adv Health Econ Health Serv Res, Vol. 19, pp. 43-69.<br />

Shojania, K.G. et al., (2010) "Effect of point-of-care computer reminders on physician behaviour: a systematic review."<br />

CMAJ, Vol 182, No.5, 23 Mar, pp 216-225. Kawamoto K, Houlihan CA, Balas EA,<br />

Sprague, R.H. (1980) "A Framework for the Development of Decision Support Systems." Management Information<br />

Systems Quarterly, Vol. 4, No.4, pp 1-26.<br />

126


Factors Determining the Adoption of ICTs in Healthcare<br />

Service Delivery: A Developing Country Context<br />

Taurai Chikotie, Jonathan Oni and Vesper Owei<br />

Cape Peninsula University of Technology, Cape Town, South Africa<br />

chikotiet@cput.ac.za<br />

jo45ng@yahoo.com<br />

OweiV@cput.ac.za<br />

Abstract: The proliferation of Information and Communication Technologies (ICTs) in healthcare service delivery has<br />

brought with it great advancements in this sector. This has created the concept of eHealth which is a relatively new<br />

concept in health care service delivery in most developing countries. According to previous research studies, ICT<br />

solutions are rapidly changing the way health organizations and stakeholders communicate with each other (Ouma<br />

and Herselman, 2008). Many people are using such communication technologies to communicate with others to<br />

gather information, with healthcare being one of the most important areas of such applications. It is therefore<br />

understandable that the value and importance of ICTs in health has increasingly been recognized world wide. The<br />

literature in innovation adoption notes that there are still inadequacies in identifying the factors that are crucial in<br />

promoting eHealth adoption in developing countries (Drury, 2005). Although, in Africa, the advent of eHealth has<br />

brought exciting opportunities to reduce or control the growing healthcare inequities, a lot still needs to be done in the<br />

adoption of these technologies (UNESCAP, 2007). In a developing country like South Africa, the disparity in the<br />

access to ICTs in healthcare service delivery is wide due to the dichotomous demography of first and second<br />

economies existing in this country. This has had a negative impact on the adoption eHealth technologies thus,<br />

suggesting that even though there has been an appreciation in ICTs in healthcare, very little has been done to<br />

ensure the adoption of such technologies. This paper, however attempts to ascertain such factors that determine<br />

adoption of ICTs in healthcare service delivery from both the management and patients’ perspectives. Drawing from<br />

the theories of innovation adoption in healthcare service delivery, the paper argues for the need to examine eHealth<br />

service adoption factors in developing countries and provides recommendations on how to tackle the challenges to<br />

adoption. The paper will conclude by recommending further research on issues in a number of key areas that need to<br />

be resolved to improve upon the efficient use and adoption of ICTs in healthcare service delivery amongst<br />

developing countries. South Africa is used as a case in this paper.<br />

Keywords: eHealth, challenges, theoretical models of adoption, ICTs in healthcare, developing countries<br />

1. Introduction<br />

The widespread use of new ICTs into almost every aspect of our professional and social lives is already<br />

gripping most developing countries (NTIA, 2007). Many people are using ICTs to communicate with<br />

others to gather information, with healthcare being one of the most important areas of such applications.<br />

It is therefore understandable that the value and importance of ICTs in health is increasingly being<br />

applauded world wide (Marconi, 2002). ICT adoption by various players in the health sector has been<br />

accelerating as well as spreading geographically in the recent years and this has led to the coinage of the<br />

new term ‘eHealth’ (UNESCAP, 2007).<br />

EHealth has several connotations, though all of them hover around the use of ICTs in healthcare service<br />

delivery. EHealth is defined as an emerging field in the intersection of medical informatics, public health,<br />

and business. It refers to health services and information delivered or enhanced through the Internet and<br />

related technologies (JMIR, 2008). Although in Africa the advent of eHealth has offered an exciting<br />

opportunity to reduce or control the growing healthcare inequity, a lot still needs to be done in the<br />

adoption of these technologies. In South Africa the gap in the use of ICTs to effective healthcare service<br />

delivery between the private and the public sector is still wide and this has had a negative impact on the<br />

adoption and use of ICTs in the healthcare sector (Bassett, 2000). This suggests that even though there<br />

has been an appreciation in ICTs in healthcare, very little has been done to ensure the diffusion and<br />

adoption of such technologies in this country.<br />

In developing countries, ICTs have been crucial to the development of effective health information<br />

sharing and dissemination systems, with users’ adoption of ICTs being one of the most important<br />

success factors of eHealth projects (Pannarunothai and Speedie, 2001). Many eHealth studies have<br />

been carried in Africa we however note that most of it has been focusing on particular technologies and<br />

only a few on the socio-technical and economic factors affecting the pace of adoption of such<br />

technologies in the healthcare sector. In modern trends of Information and Communication Technology, it<br />

is difficult for a project to be implemented without the consideration of personal, institutional and national<br />

127


Taurai Chikotie et al.<br />

factors especially in the adoption of new technologies (Ankem, 2004). This paper therefore outlines the<br />

research question as: What factors determine the adoption of ICTs in the healthcare service delivery<br />

sector in South Africa? The main purpose of this paper is to identify those factors that are crucial to the<br />

adoption of ICT in the healthcare delivery sector in South Africa.<br />

The rest of the paper is structured as follows: The next section is a literature review of healthcare and ICT<br />

development. This is followed by theoretical models of adoption and healthcare service delivery in section<br />

3. Section 4 discusses our research methods. An analysis of our findings is presented in section 5. Also<br />

in this section, the findings are discussed. The paper concludes in section 6 with a summary and a<br />

discussion of eHealth service delivery in developing countries.<br />

2. Healthcare and ICT development<br />

2.1 Global ICT developments in healthcare service delivery<br />

In recent years, the rates of adoption and diffusion of ICTs have become important indicators of<br />

development. Developing countries have put technological innovation on top of their national<br />

development agenda and invested heavily in ICTs (Detmer, 2000). For example, in the healthcare sector,<br />

the adoption of ICTs is perceived as the main driving force in the unfolding healthcare reforms in many<br />

developed and developing countries (Gladwin, Dixon, and Wilson, 2003). The influx of ICT in healthcare<br />

has increasingly been recognized by government agencies, health practitioners, private sector, academia<br />

and NGOs across the world as a milestone in quality health delivery. The United Nations Economic and<br />

Social Commission for Asia and the Pacific (UNESCAP) report on ICT for development (2007) stated<br />

that, judging from current news, research and reports, ICTs adoption by various players in the health<br />

sector has been accelerating as well as spreading geographically in the recent years and this is<br />

confirmed by the number of eHealth initiatives that have sprung up spreading from metropolitan areas to<br />

remote rural areas, increasingly, in both developed and developing countries. Wallace (1997:88) notes<br />

that in today’s information society, health professionals need to maximize the potential benefits offered by<br />

ICT as a means to improving public access to health care and information. Globally the convergence of<br />

ICTs has changed the way in which the health sector and society at large are organised and managed.<br />

From the foregoing it is seen that, it is important for every part of the health sector to adapt to these<br />

changes as ICTs can enhance the delivery of health services in various ways.<br />

2.2 ICT development in healthcare service delivery: South Africa<br />

Citizens in most of the emerging economies are beginning to taste the positives of communication<br />

technologies especially in the way they have improved the quality of their health lifestyle (United Nations,<br />

2008). In an era in which the state of health for developing countries, communities, and people in the<br />

world is at best stagnant and is probably in deterioration, the potential value of new information<br />

technologies for more effective health and development action can appear to be manna from heaven<br />

(Feek and Long, 2003:2). Emerging and existing health and ICTs initiatives have been put in place for the<br />

benefit of communities and also exposing key issues that need addressing in future projects.<br />

Most governments in developing countries are appreciating the critical role of ICTs to change their<br />

countries and have started to adopt these ICTs for national development. A lot of transformations in the<br />

way governments operate have been brought about by the advent of ICTs (Scott et al, 2005:97). Enakrire<br />

and Onyenania (2007:15) allude that most African countries have acknowledged the information<br />

revolution, although the adoption of communication technologies differs from country to country.<br />

The South African government has emphasized the development of an ICT sector through the formation<br />

of a national ICT strategy that addresses ICTs penetration and adoption particularly in underserved<br />

communities.<br />

According to the Department of Industry and Trade (2007), there has been the formulation of the South<br />

African Information Technology Industry Strategy (SAITIS) which is a bilateral project between the South<br />

African government and the Canadian government. This initiative seeks to contribute to sustainable<br />

economic growth, social upliftment and empowerment. Apart from SAITIS and Info 2025 Vision that have<br />

been assigned responsibility of building ICTs infrastructure, three taskforces were put in place to address<br />

to address ICTs deployment as a socio-economic development enabler and these are:<br />

128


Taurai Chikotie et al.<br />

The Presidential International Taskforce on information Society, which targets mainly the global ICTs<br />

market;<br />

The National Information Technology Taskforce, which deals with local ICTs initiatives and;<br />

The IT Council, responsible for local and provisional government information technology functions<br />

(Digital Opportunity Initiative, 2001: 13).<br />

There have also been a number of ICT initiatives in South Africa focusing on equitable information<br />

access and dissemination in the healthcare service delivery sector. Such initiatives include the Project<br />

Masiluleke (BBC, 2008); the telemedicine project done by the Medical Research Council and the<br />

Department of Health (DoH, 2007); the LoveLife Project (Chetty, 2007) and the Closed Health Broadcast<br />

Channel (Feek and Long, 2003). The main objective of such initiatives was to tap the power of ICTs as a<br />

high impact, low cost and efficient tools to deliver healthcare services at a distance to South African<br />

communities. They also seek to provide underserved communities with access to quality health<br />

information through the use of information communication technologies. However, research studies<br />

suggest that South Africa is poorly utilising Information and Communication Technologies (ICTs) in<br />

healthcare in disadvantaged communities although it is the very country that is leading in ICT<br />

development in the African Continent (Research ICT Africa, 2008; Gillward et al, 2005).<br />

3. Theoretical models of adoption and healthcare service delivery<br />

Several theoretical models were discussed to gather information on factors that foster the adoption of<br />

innovation from all perspectives, that is, social, technical and institutional or managerial. Such factors are<br />

deemed crucial in the adoption of ICTs within the healthcare sector from both the patient and the<br />

management or institution’s point of view. The most widely used of these theoretical models are the<br />

Diffusion of Innovation (DOI) which seeks to identify how an innovation is communicated through certain<br />

channels over time among the members of a social system (Rogers,1995); the Social Learning Theory<br />

which considers that people learn from one another, including such concepts as observational learning,<br />

imitation, and modeling (Bandura,1977); the technology acceptance model (TAM) which uses<br />

perceptions to predict attitude towards using or rejecting a technology or innovation (Davis, 1989); and<br />

the Fit between Individuals, Task and Technology (FITT) which seek to use the interaction of users, task<br />

and technology as the basis to better understanding IT adoptions in healthcare environments<br />

(Ammenwerth, Iller and Cornelia ,2006). All of the above mentioned theories seek to argue the adoption<br />

of innovations from the technology’s point of view and not much from the institutional or management’s<br />

perspective though they do address some of the social concerns. The identification of various factors on<br />

the management and patient view in South Africa is examined and proposed in the development of a<br />

pragmatic framework. Figure 1 depicts the proposed framework that will further be used in the findings<br />

from survey questions and interview.<br />

4. Experiment<br />

4.1 Research methods<br />

Document analysis: Extensive academic literatures on eHealth development and community informatics<br />

(Mbarika, 2007; Eysenbach, 2008; Samake and Mbarika, 2004; and Gurstein, 1999), ICT policies in<br />

healthcare (Rodrigues, 2008 and World Health Organisation, 2008), challenges to the adoption of ICT in<br />

healthcare (Kiplang’at & Ocholla, 2005; Migiro, 2006) were reviewed; including journals on previous<br />

related research innovations adoption in healthcare (Bates, Manuel and Oppenheim, 2007) and<br />

(Borgatti,2005), as well as white papers. These articles were analyzed and the gaps in ICTs adoption in<br />

healthcare were identified. The objective was to ascertain the factors critical in the assimilation of ICTs in<br />

the healthcare sector in a developing country scenario.<br />

For the research survey and interviews, patients and healthcare personnel from two healthcare centres<br />

(both public and private) in the City of Cape Town participated.<br />

Questionnaires: Survey questionnaires were issued to 20 patients in which they had to respond to<br />

questions on their use of ICTs of any type in health information communication, their frequency of use of<br />

such ICTs, the benefits they have gained and the challenges the are facing.<br />

129


Taurai Chikotie et al.<br />

Figure 1: Proposed framework for eHealth service delivery<br />

Interviews: Face to face interviews were conducted with 5 senior healthcare personnel with an<br />

understanding of ICTs within their organizations. These were used as the key informants to complement<br />

most of the data from questionnaires.<br />

The consultants were asked about their specific experiences working with ICTs and what they found<br />

when working with such technologies. Most of the responses focused on their brief, internal and external<br />

healthcare projects (i.e. benefits, challenges, strategies or policies etc).<br />

5. Analysis of findings<br />

Our findings identified several critical factors that are instrumental in the adoption of ICTs in healthcare<br />

service delivery most notably within a developing country context. The responses of the interviews and<br />

questionnaires are presented and discussed from both the management and patient’s perspectives. Most<br />

of these responses are closely linked to the understanding of ICTs usages and the challenges being<br />

faced in using these technologies. Important to note is the fact that while there has been several ICTs<br />

and healthcare related projects in developing countries, such projects have not been able to sustain<br />

national needs with some failing to meet their objectives as their models or frameworks are based on<br />

alien foundations different from what developing countries need.<br />

Several factors were identified from the interviews with senior healthcare personnel and these<br />

include:<br />

Knowledge: Most of the senior healthcare personnel concurred to having serious challenges on issue<br />

such as lack of end-user training on the systems they use, lack of online health resources awareness and<br />

poor research skills in finding authoritative information online. This was confirmed by another interviewee<br />

who noted that their academic background is to blame for the slackness in ICT skills.<br />

Organizational: some respondents brought issues such as the need to satisfy diverse stakeholders, the<br />

need for timely response to service requests, resistance to change, and the need to integrate services<br />

with the workflow of medical, technical and administrative staff.<br />

130


Taurai Chikotie et al.<br />

Regulatory and policy: these include the lack of concrete security standards for medical records, Issues<br />

of interstate licensure, credentialing of health care professionals and liability for malpractice, and lack of<br />

suitable payment mechanisms. The healthcare sector is a very sensitive field which needs stern<br />

measures in the case of information sharing and dissemination. Just like in developed countries, the<br />

issue of ethics and regulations play a crucial role in the use of ICTs in healthcare service delivery.<br />

Economic: includes the lack of sustainable funding models, and pace with the ever-changing technology<br />

while controlling costs. In South Africa the healthcare sector is always operating at shortfall due to low<br />

funds being allocated to the department. This leaves most of the management to weigh priorities for<br />

healthcare service delivery, that is, it is either drugs or technology of which the latter always play second<br />

fiddle.<br />

From the patient’s perspective the following factors were identified:<br />

Poor user awareness on new technologies: in spite of the support for the exciting benefits of eHealth, a<br />

number of obstacles continue to stand in the way of its widespread adoption by health organizations and<br />

consumers. Most consumers are still not aware that they may access specialist knowledge online (Kedar<br />

et al, 2003). Most patients link this to poor awareness on the abilities of ICTs especially in health<br />

information sharing and dissemination.<br />

ICT Infrastructures: South Africa, like any other developing country, still faces a lot of challenges in ICT<br />

infrastructural development. This contributes significantly to the poor adoption of eHealth in South Africa.<br />

Many African countries with South Africa included have witnessed considerable local and foreign<br />

investments in ICT infrastructures development, primarily in the telecommunications sector and it is up to<br />

the policy makers to take this as an opportunity to enhance the use of ICTs in the health sector.<br />

Education: most responses also implied that both well educated health staffs and patients have positive<br />

impacts on the development, implementation, adoption, and diffusion of various eHealth initiatives.<br />

Tornatzky and Klein (1982) write that, an effective education facilitates and increases the expression and<br />

identification of patients’ needs, the communication between providers and patients, and the patients’<br />

awareness and appreciation of the benefits of the eHealth project. Incorporating ICTs into medical<br />

educational interventions is a challenge. It is therefore vital that staff is trained on ICTs use to ensure<br />

maximum effectiveness and efficiency for support in the process of rendering health services to patients<br />

(Sargeant, 2007).<br />

Content divide: information provided on most ICTs is in English, which is also a hindrance to non-English<br />

speaking people who might need to understand the whole concept coming with these health<br />

technologies, and South Africa is not an exception (Mbarika and Kifle, 2006). This is a concern to most<br />

patients especially those who use mobile phones and have a poor English background.<br />

6. Discussion and conclusion<br />

6.1 Summary<br />

From the findings presented in this paper, it is evident that adoption of ICTs in healthcare service delivery<br />

from both the management and patients’ perspectives comes with various challenges. Drawing from the<br />

theories of innovation adoption in healthcare service delivery, the paper argues for the need to examine<br />

eHealth service adoption factors in developing countries. On the patience perspective several issues like<br />

poor user awareness on new technologies, education, content divide, etc. has been identified. These<br />

factors will go a long way in assisting policy makers to evaluate the need and effectiveness of healthcare<br />

service delivery especially in a developing country context. However, senior management in healthcare<br />

will especially play an important role in sharing and dissemination of eHealth policies to patients. Just like<br />

in developed countries, the issue of ethics and regulations play a crucial role in the use of ICTs in<br />

healthcare service delivery.<br />

6.2 Government policies<br />

The lack of clear government policies and strategies in the promotion of eHealth in most developing<br />

countries has led to the slow uptake of the technologies. Where initiatives have been implemented, there<br />

has not been adequate monitoring and evaluation as to how these initiatives are performing (Ouma, and<br />

131


Taurai Chikotie et al.<br />

Herselman, 2008). It is important that government policies support most of the initiatives to promote<br />

eHealth.<br />

Consequently, some of these factors may differ from one community to another and may not directly<br />

reflect a particular eHealth environment. But it is hoped that this study would be applied to various<br />

organizations, governments, and eHealth policy practitioners in planning eHealth adoption. This paper<br />

will also be of practical relevance in the decision making process of senior management and consultants<br />

of ICTs in service delivery. The various factors identified as well as findings from interviews in a<br />

developing world context will address and further highlight the need to promote eHealth especially in the<br />

developing countries.<br />

References<br />

Ammenwerth, E, Iller, C and Corneila, M. 2006. IT-adoption and the interaction of task, technology and individuals: a<br />

fit framework and a case study. BMC Medical Informatics and Decision Making. Vol 6(3)<br />

Ankem, K. 2004. Use of Internet resource-based value-added processes by faculty in LIS education. Library and<br />

Information Science Research, 28:482-500<br />

Bandura, A. 1977. Social Learning Theory. Englewood Cliffs, NJ: Prentice-Hall.<br />

Bassett, H. 2000. Healthcare in South Africa. MedHunters Articles. URL:<br />

http://medhunters.com/healthcareInSouthAfrica.html. (Online) Retrieved: 07 September 2010.<br />

Bates, M., Manuel, S. & Oppenheim, C. 2007. Models of early use of ICT innovations in higher education. ARIADNE,<br />

50:1-10.<br />

Borgatti, S. P. 2005. Thinking theoretically. URL: http://www.analytictech.com/mb870/handouts/theorizing.htm :<br />

[Accessed: 8 October 2010].<br />

Chetty, K.S.2007. EHealth: the South African Context. URL: www.doh.gov.za/healthini.htm . [Accessed: 14<br />

November 2010]<br />

Department of Health. 2007. Health Strategic Framework 1999-2004. URL: http://www.doh.gov.za/docs/policy-f.html.<br />

[Accessed: 23 March 2010]<br />

Department of Trade and Industry, 2004. Benchmarking of technology trends and technology developments report.<br />

Johannesaburg: Bluepeter/AMI.<br />

Detmer, D.E. 2000. Information Technology for Quality Health Care: A Summary of United Kingdom and United<br />

States Experiences. Quality in Health Care, 9: p. 181-189.<br />

Digital Opportunity Initiative, 2001. Creating a development dynamic: Final report of the Digital Opportunity Initiative.<br />

Okinawa: UNDP<br />

Drury, P. 2005. The eHealth agenda for developing countries. World Hosp Health Serv; 41: 38–40.<br />

Enakrire, T. R. and Onyenania, O. G. 2007. Factors affecting the development of information infrastructure in Africa.<br />

Library Hi Tech news, 2: 15-20.<br />

Eysenbach, G. 2001. What is eHealth?. Journal of Medical Internet Research 3(1):e20. URL:<br />

http://www.jmir.org/2001/2/e20/index.htm. (Online) Retrieved: 3 October 2010.<br />

Feek, W and Long, G. 2003. The Digital Pulse: The Current and Future Applications of Information and<br />

Communication technologies for developmental Health Priorities. A Report Published by the Communication<br />

Initiative. URL: www.comminit.com [Accessed: 7 January 2010].<br />

Gillwald, A. and Esselaar, S. 2005. South African ICT Sector Performance Review. Wits University.School of Public<br />

and Development Management, Johannesburg.<br />

Gladwin, J., Dixon R.A., and Wilson T.D. 2003. Implementing a new health management information system in<br />

Uganda. Health Policy and Planning. 18(2): p. 214-224.<br />

Gurstein, M and Dienes, B.1999. “A ‘Community Informatics’ Approach to Health Care for Rural Africa”. Presented to<br />

The Africa Telemedicine Project: <strong>Conference</strong> ’99<br />

Jai Ganesh, A.U. 2004. EHealth - drivers, applications, challenges ahead and strategies: a conceptual framework.<br />

International Journal of Medical Informatics. Vol (20)2<br />

JMIR. What is eHealth? Journal of Medical and Internet Research. 3(1): e20. URL:<br />

http://www.jmir.org/2001/2/index.htm. (Online) Retrieved : 8 November 2010<br />

Kedar, I et al. 2003. Internet based consultations to transfer knowledge for patients requiring specialised care. British<br />

Medical Journal. Vol (326). 696-699<br />

Kiplang’at, J. & Ocholla, D. N. 2005. Diffusion of information and communication technologies in communication of<br />

agricultural information among agricultural researchers and extension workers in Kenya. South African Journal<br />

of Libraries and Information Science, 71(3):234-246.<br />

Marconi, D. 2002.Health Connectivity for development in America. Social Science and Medicine. University of<br />

Michigan: University Press. 40-43<br />

Marcus, J.A. 1986. Diffusion of innovations and social learning theory: Adoption of the context text-processing<br />

system at Stanford University. Dissertation Abstracts International, A 46/12, 3553(University Microfilms No. AAT<br />

8602506)<br />

Mbarika, V and Kifle, M.. 2006. Telemedicine in sub-Saharan Africa: The case of tele-ophthalmology and eye care in<br />

Ethiopia. Journal of the American Society for Information Science and Technology. 57(10), 1383 -1393<br />

Migiro, S. O. 2006. Diffusion of ICTs and e-commerce in manufacturing SMEs in Kenya. South African Journal of<br />

Library and Information Science, 72(1):35-44.<br />

132


Taurai Chikotie et al.<br />

NTIA (National Telecommunications and Information Adminstration). 2007. Falling through the Net: Towards digital<br />

inclusion. Department of Commerce: Washington, DC. 47-53<br />

Ouma, S and Herselman, M.E. 2008. EHealth in rura areas: case study of developing countries. Proceedings of<br />

world academy of science, engineering and technology. Vol (30) 560-566<br />

Pannarunothai, S., Speedie S. 2001. Factors influencing health information technology adoption in Thailand&apos;s<br />

community health centers: Applying the UTAUT model. International Journal of Medical Informatics, Vol (78)6.<br />

404-416<br />

Research ICT Africa .2008. ICT Access and Usage in South Africa, LINK Centre. University of Wits, Johannesburg,<br />

SA.<br />

Rodrigues, R.J. 2008. EHealth Strategies in America. The Commonwealth Health Ministries Reference Book. 5(3):<br />

e4. URL: http://www.ehealthstrategies.com .[Accessed: 19 June 2010]<br />

Rogers, E.M. 1995. Diffusion of Innovations. 4 th Ed. New York, N.Y.: Free Press. 58-63<br />

Samake, K.D. and Mbarika, W.A. 2004. EHealth in Africa: A vision for a healthier Africa. American Journal of Public<br />

Health. 96(1). 73-78<br />

Sargeant, J.M. 2007. Medical education for rural areas: Opportunities and challenges for information and<br />

communication technologies. Journal of Postgraduate Medicine. 51(4): 301-307<br />

Scott, V., Chopra, M., Conrad, E and Ntuli, A. 2005. Low equitable is the scaling up of HIV service provision in South<br />

Afnca? SAfrMedf. 95:I09-n3.<br />

Tornatzky, L and Klein, K. 1982. Innovation characteristics and innovation adoption- implementation: a meta analysis<br />

of findings. IEEE transactions on Engineering Management. Vol. 29(1), 28-48<br />

[6] UNESCAP. 2007. ICT for development in South East Asia and the Pacific. Report on developing countries and<br />

ICT revolution for the Scientific and Technological Options Assessment Panel. URL:<br />

www.unescap.org/search/rpts/ictdcp_asia466878764.pdf. (Online) Retrieved: 24 September 2010<br />

United Nations Foundation.2009.Global ICTs Report 2009.<br />

World Health Organisation. 2008. ICT and healthcare in the African region. URL:<br />

http://www.who.org/search/africa_sa.htm. [Accessed: 27 October 2010]<br />

Websites: http://newsvote.bbc.co.uk/2/hi/technology/7688268.stm. (Online) Retrieved: 11 October 2010.<br />

133


Evaluation of Application Embedded Knowledge Migration<br />

Issues<br />

Mitchell Cochran<br />

Claremont Graduate University, Claremont, USA<br />

mcochran@ci.monrovia.ca.us<br />

Abstract: As computing has matured, more organizations are purchasing best of breed applications as opposed to<br />

developing them in-house. From a Knowledge Management point of view, the organizations are renting the use of<br />

knowledge that is embedded in the applications. The organizations may own the data but the host application<br />

company owns the intellectual capital that creates the knowledge. For any of a number of reasons organizations will<br />

have to move to new applications and in turn new knowledge. It is assumed that the organization will be able to<br />

migrate current data and print reports but it does not own the original base knowledge. The issue is to understand<br />

what knowledge is imbedded in the old application and how can it be integrated into the new system. As the<br />

knowledge is inventoried, the new vendor can then determine if the knowledge will be available in the system. After<br />

that determination, the user might have to decide if the information is obsolete or possibly lost data. The migration<br />

issue also can put the onus of development on the end user. Consider the conversation of the developer and the end<br />

user where the end user asks for a feature that the developer has not seen. The end user is looking for features in<br />

the old system and the developer is going to say that it is up to the end user to tell them what they want. The issue is<br />

that the end user may now know what they want. The knowledge embedded in the code of the old application<br />

provided the information. The information basis is the intellectual property of the outgoing vendor and they may not<br />

have any reason to work with the incoming vendor. The paper will evaluate migration issues based on a case study<br />

of the migration of a financial application for a small city. The paper will also discuss some of the assumptions of<br />

knowledge management and a knowledge inventory to help an organization prepare to move applications to a new<br />

vendor.<br />

Keywords: knowledge management, application, migration<br />

1. Introduction<br />

An organization is responsible to understand all of the knowledge which is necessary to run its business.<br />

The knowledge is the intellectual capital of the organization. Rastogi defines intellectual capital as the<br />

“holistic meta-level capability of an organization to generate creative and effective responses to extant<br />

and emerging, present and potential challenges facing it, in an ongoing manner”. (Rastogi, 2000)<br />

The firm may or may not have developed explicit practices to maintain the knowledge base. Instead, it<br />

may be dependent on informal practices. Nelson and Winter comment that much of an organization’s<br />

knowledge is tacit because the crucial know-how resides in the minds of the organization’s members.<br />

(Nelson and Winter, 1980) Badaracco comments that embedded knowledge tends to be non-migratory<br />

where it “resides primarily in specialized relationships among individuals and groups and in the particular<br />

norms, attitudes, information flows and ways of making decisions that shape their dealings with each<br />

other.” (Badaracco, 1991) The organization needs to develop a knowledge inventory to develop an<br />

understanding of the critical process or application knowledge. Without a formal procedure to identify the<br />

knowledge inventory the company may not be able to fully document, develop or migrate knowledge to<br />

new applications.<br />

Many authors have identified the importance of documenting knowledge. Bots and deBruin created a<br />

knowledge value chain which includes a ‘Make Inventory Determine Knowledge Needed’. (Bots and<br />

deBruin 2002). Sage and Rouse identified issues that also relate to knowledge identification and<br />

migration: (Sage and Rouse 1999)<br />

Modeling processes to identify knowledge needs and sources<br />

KMS strategy for the identification of knowledge to capture and use and who will use it<br />

An understood enterprise knowledge structure<br />

Buhler states that only developing learning organizations that use knowledge that they acquire can<br />

continue to adapt and respond to their changing environment (Buhler 2002). McManus et al summarize<br />

the challenge to managers: ‘Managers must adequately plan for and implement KM practices that will<br />

allow them to effectively operate despite the multifarious influences that threaten the knowledge bases of<br />

their business.’ (McManus et al, 2003) The KM practice that this paper focuses on is developing an<br />

inventory of knowledge.<br />

134


Mitchell Cochran<br />

The importance is not which tool is used to define the knowledge but that the organization defines and<br />

follows a plan to document the knowledge. There any number of tools that could be used. Grossman et al<br />

list four different KM models: the Balanced Scorecard, The Intangible Assets Monitor, the Skandia<br />

Navigator, and the Intellectual Capital Index. (Grossman et al, 2006). If the organization is not using a<br />

tool, they may not have a KM strategy in place for either the functional business area or the overall<br />

organization. In 2010, an informal survey of municipal information system directors for the State of<br />

California asked how many agencies had complete documentation of all of their systems. Approximately<br />

10% of the 300 cities or special districts responded that they had completely documented their systems.<br />

The survey was only intended to get an initial impression of the issue’s scope. The entities had detailed<br />

documentation on which applications were in use and some went so far as to have documented the<br />

system settings. Most have only created electronic backups without a way to recreate the application<br />

settings should it need to be reinstalled from the vendor supplied media. The backups can be applied to<br />

new equipment but the organizational knowledge within may not be understood. The systems have been<br />

documented but not the embedded knowledge.<br />

If there is a plan or procedure, an implied part of the plan is to be able to identify all of the knowledge<br />

used by the firm. The original sources of knowledge may be internally or externally developed such as<br />

outsourcing. In either case, the components of the organization’s functional knowledge may readily<br />

available. Externally provided knowledge may be obtained by purchasing applications, outsourcing<br />

services, or using consultants. However, the knowledge may be tacit in nature so that it can not be easily<br />

recreated or available in a raw form. The knowledge may be available from the consultants or have to be<br />

purchased for a second time. Internal knowledge may have been developed by sources that are no<br />

longer with the firm which can be the case for retiring employees. The foundations and assumptions that<br />

created the legacy knowledge may not be documented.<br />

One of the critical areas would be for the firm to understand what processes are known and what are<br />

unknown. The knowledge inventory would help to identify what knowledge points are known and which<br />

attributes need to be discovered. Without a process to indentify what the organization knows, it cannot<br />

identify what points are missing or unidentified.<br />

Without a detailed understanding of the knowledge base the firm may not be able to recreate it in other<br />

forms. The issue is that the firm may not understand how to move the company’s knowledge base to new<br />

applications or processes. From a disaster recovery context, many organizations may not understand the<br />

knowledge inventory which is required to operate an organization without any automated processes.<br />

Many organizations do not map all of the information that is used by the organization. They may map<br />

what they consider as the critical missions. One of the concerns is that not all of the knowledge is<br />

identified. The critical base of knowledge may be dependent on processes that are not documented. The<br />

organization needs to identify and understand all of the processes whether the processes are internal or<br />

external, and or known or unknown.<br />

As part of a knowledge inventory the organization needs to understand which knowledge it controls and<br />

knowledge that is controlled by an outside entity. Some of that knowledge is obtained from consultants<br />

and available to be inventoried. The consultant will typically deliver a product or report that documents<br />

the results of the services provided. Some of the knowledge base will not be available since it was<br />

purchased as an application. The application takes inputs, applies an embedded knowledge and then<br />

develops an output. The process knowledge may not be known. The software will probably include a use<br />

license as opposed to an ownership agreement which might infer a right to re-engineer the application.<br />

Many applications with embedded knowledge can not be re-engineered due to those legal or contractual<br />

agreements.<br />

Should a firm use outsourcing services, the firm may not own the knowledge of the application, just the<br />

data used to create it. The organization will need to understand the process if the organization plans on<br />

providing the capability internally in the future. This situation can be found in hosted applications or many<br />

cloud computing environments. A simple example would be a outsource firm that provides payroll<br />

processing. The firm provides the knowledge of current personnel benefit, tax and withholding issues.<br />

An organization needs to have a method to document the organization’s knowledge. A framework or<br />

process needs to identify what knowledge is required and or which knowledge is readily available. If the<br />

data or knowledge is not available then there will be a cost for its recreation. One example is to identify<br />

135


Mitchell Cochran<br />

the level of knowledge that the organization requires about human resource practices. The organization<br />

can easily outsource HR practices to industry leading firms. The knowledge that is easily found in the<br />

industry can be easily migrated. The more specialized the information is the more difficult it will be to find<br />

outside vendors to provide it. The organization needs to understand if that knowledge is strategic to the<br />

key practices of the organization. The organization’s knowledge base will include organizational memory<br />

theory to develop the existing structure of retained knowledge.<br />

Many forms of knowledge will have a shelf life. The organization will have to determine what the cost<br />

would be to recreate any data that is not maintained. One example is the data on an organization’s<br />

infrastructure. If the information updates are not applied to the knowledge base then the knowledge is<br />

lost. All that will be available will be the obsolete data. Consider what it would take to recreate the US<br />

space program so that a person could be landed on the Moon in the near future. The technology has<br />

advanced over 30 years but the operational knowledge of the engineers was lost due to personnel<br />

retirements. That operational knowledge would have to be recreated. (Leonard, Kiron, 2002)<br />

Another concern is to understand if a process is outsourced, what would be the cost to bring it back into<br />

the organization. Some operations may be outsourced due to cost savings. If the outsourcing is not as<br />

effective as desired or if the outsourced operation has become more integrated into the critical mission it<br />

may need to be brought back into the organization. The initial operational knowledge may have to be<br />

recreated or purchased.<br />

Unidentified or missing knowledge will also be an issue. Over time, the data origins or the knowledge<br />

itself may be unavailable. Tacit knowledge may leave as employees leave. Legacy applications may<br />

contain knowledge but not the foundations that created that knowledge. An organization may understand<br />

the processes or rules from legacy applications but they need to understand the assumptions and<br />

theories needed to create those processes. The organization needs to answer the question of why it did<br />

what it did. As the environment changes, the organization will have to adapt the rules. If it does not<br />

understand the knowledge behind the rules or logic then it will be costly to alter the logic. If it is not<br />

documented, the knowledge will only be retained by the memory of the staff. The issue of knowledge<br />

retention is especially important with the retirement of the experienced baby boomer workers. A simple<br />

example can be considered when deciding if an accounting operation is taxable. There are rules that<br />

govern the process but how the organization handles an exception will require documentation of the<br />

assumptions or circumstances. The rules try to create black and white situations but there will always be<br />

a grey area. The answers to how and why the organization achieved the result are just as important as<br />

the result. Without the assumptions or background information the operation results may not be<br />

repeatable.<br />

For disaster recovery purposes, company needs to be run an operation without its primary tools. Should<br />

the organization not have access to its applications, it would have to recreate its processes using other<br />

procedures or manual tools.<br />

Many knowledge concepts are embedded in applications. The knowledge is made tacit by the<br />

programmer in the application. The issue is that the knowledge is implicit to the application and may not<br />

be available to the user. The user can own the data and use the application to generate results but he or<br />

she may not understand the knowledge area. One simple example would a tax application. The rules are<br />

documented and available but the user does not want to spend the resources to understand the<br />

application area. The user would rather purchase the application where the domain knowledge is then<br />

made accessible. An interesting example would be how the user would respond to a tax audit based on<br />

the results of the application. The user can only attest to the data that was input in the application. The<br />

user cannot attest to the appropriateness of the results, only a review of the application logic can provide<br />

that support.<br />

The organization’s application base provides knowledge of the firm’s operations. The organization will<br />

optimize procedures based on the applications. Over time, those applications will change due to vendor<br />

requirements, hardware obsolescence, or new customer requirements. If the current application process<br />

and procedures are not understood then it will be difficult to migrate to a new application. The current<br />

vendor owns the application and has no incentive to help an organization move to a new application. The<br />

user has ownership of the data but not the knowledge. Consider the following conversation between a<br />

customer and vendor:<br />

136


Mitchell Cochran<br />

Customer: Please provide a similar function that we had in the old system.<br />

Vendor: How did the function work?<br />

Customer: We don’t know, the program provided the information.<br />

Vendor: As a vendor we can’t re-engineer the old code so we don’t know what it did or how it did<br />

it. You will have to tell us what you want and how to do it.<br />

Customer: But we don’t know how it works.<br />

The current vendor may not be able to provide technical assistance or a description of the knowledge that<br />

the application contains. The new vendor asks the customer what they would like. The customer<br />

responds that they want the application to perform like the last one. The new vendor may not have an<br />

understanding of the old application. The onus will be on the customer to develop a knowledge base of<br />

what the old application did and what the new application should do. If the customer does not understand<br />

the knowledge base they will have to incur the cost to recreate the knowledge.<br />

2. Case study<br />

The City of Monrovia, California was required to upgrade its Cayenta financial application during 2007.<br />

The vendor was no longer going to support the installed version. The City would have been content to<br />

continue with installed version but the change was required by the vendor’s decision to terminate support.<br />

The new version included significant new features and would cost $500,000 for both the software license<br />

fee and migration efforts. Due to the large dollar amount, the City was required to investigate alternatives.<br />

The Request for Proposal or RFP process had eight proposals which ranged in cost from $300,000 to<br />

$1,000,000. The City chose a new vendor, Munis, which bid approximately $350,000 for the software<br />

license fees and migration costs. The evaluation did not show enough value in the $500,000 Cayenta<br />

proposal to justify the additional $150,000 expense to stay with the original vendor. The new vendor only<br />

had one similar financial system installed in California. There are special reporting mechanisms for the<br />

California State Public Employee Retirement System or PERS. Each vendor needs to develop the<br />

custom reports. The new vendor promised to develop the required reporting code in a timely manner.<br />

MUNIS had similar reporting requirements for other states so they assumed that the level of effort would<br />

be comparable. The California cities of Lakewood, Palmdale, Redondo Beach and Chino also purchased<br />

the financial system approximately at the same time. The cities joined together into a local users group.<br />

All of the cities had the same retirement system reporting requirements.<br />

The previous vendor had a working retirement reporting system. The new vendor had yet to develop its<br />

system. The new vendor required the users group to determine the needs for the new system. The users<br />

group acted as the development team. The new system vendor would develop prototypes and the users<br />

group was required to troubleshoot each version. The MUNIS developers underestimated the level of<br />

effort required. The development project took over two years. Some of the Cities had working systems,<br />

others where not completed as of this point. One of the development difficulties is that each city had<br />

different requirements. Fire and police departments have unique retirement reporting requirements.<br />

Some cities did not have public safety components which simplified their reporting efforts. The cities<br />

without either fire or police departments where able to implement a correctly working reporting system.<br />

The issue is that the vendor would commit the resources to develop the system but the responsibility for<br />

a correct submission lies with the City. If the system is delayed, which might be expected on a new<br />

system, the City has to find a work around or a manual method. The City of Monrovia was forced to take<br />

output from the new financial system and enter the batch totals into the old system to get the correct<br />

submission information. In addition to any normal or expected migration efforts, the PERS system<br />

announced that it was going to change the reporting requirements and structure. The new system was<br />

then put on hold pending the State implementation of the new requirements. The project was further<br />

complicated by delays in the State implementation of the new standards.<br />

The ultimate responsibility for submission of correct data is that of the end customer which in this case is<br />

the City. The vendor committed to develop the system but it would need specific requirements. The<br />

contractual liability of the vendor is typically limited to the amount of the fees paid by the client. The City<br />

has been forced to maintain the old system to act as a migration tool for creating the correct reports.<br />

Even with known requirements from the retirement authority, the vendor was not able to develop the<br />

137


Mitchell Cochran<br />

reporting system in a timely manner. They could develop a system that works for some entities but not all<br />

of the California clients. The issue is the complexity of integrating the new reporting requirements into the<br />

existing system. There are two parts of knowledge that need to be considered. The first is the<br />

requirements that can be developed or presented by the customer. The second part is how to integrate<br />

the new requirements into the existing system. The structure of the existing program may not be capable<br />

of handling new information. The application may not have internal structures that can support the new<br />

requirements. Knowledge management theory which discusses lost or missing data can add some insight<br />

into the migration issue.<br />

The knowledge migration requirements have not been met in trying to implement the new system.<br />

3. Knowledge inventory<br />

Most organizations collect their knowledge in a written form. The policies and procedures can be clearly<br />

documented. One of the main issues would be how current the information is. The line staff would have<br />

the most current information. As an example, emergency management guidelines recommend that the<br />

facilities managers be the first people contacted in a facilities disaster. They will have the most up to date<br />

information about the contents of the buildings and staffing. The emergency operations documentation<br />

may not be kept current.<br />

From a KM point of view, training manuals are documentation of the organization’s knowledge. Many<br />

agency training manuals take a manufacturer’s and then customize them for the local environment. The<br />

customization is one piece of the knowledge inventory. Training videos can be a tool to capture<br />

knowledge via storytelling. NASA partnered with the NASA Engineering and Safety Center (NESC) to<br />

extract knowledge from NESC experts and incorporate it into targeted training courses. (KMEdge, 2010)<br />

It is also important for the organization to understand how to create the information. Meta-data or data<br />

about the creation or use of the data is also important. Data turns into knowledge with a context. The<br />

meta-data can provide the assumptions and background to describe the context. In a Geographical<br />

Systems (GIS) Environment, the meta-data describes the mathematical model or projection that is used<br />

to describe the spatial data. If an analyst does not know or validate the projection that information cannot<br />

be combined with other spatial data. Knowledge can also be context sensitive. The basis of that<br />

knowledge is based on information and assumptions. Should the knowledge’s base information change,<br />

the current knowledge would have to be verified against the current assumptions or environment.<br />

The organization may choose to outsource some of the procedures that it does not understand, is not a<br />

cost effective user, or decides that it is not part of the core operations. For the procedures that are<br />

purchased, the knowledge may not be owned or understood. From a programmer’s viewpoint, the<br />

knowledge is embedded in the application but it is explicit since the programmer can determine the<br />

relationships and procedures. From the user’s viewpoint, the knowledge is tacit within the application.<br />

Fuller in his book, Knowledge Management Foundations, points out that programmers are converting<br />

tacit, human knowledge into explicit, transferable intellectual property. (Fuller, 2002) The knowledge is<br />

the intellectual property of the developer. The user or owner of the data may or may not have the legal<br />

ability to see the code. Many license agreements have limitations on the deconstructing or reverse<br />

engineering the application code.<br />

To develop a knowledge inventory the organization needs to address some of the following points:<br />

The necessary steps that the organization uses to identify what kept knowledge is important.<br />

The organization needs to understand the key information points that make the organization run.<br />

The organization may be a straight forward processor of information such as an accounting<br />

department which has industry standards that dictate the majority of the knowledge processes.<br />

However, the organization needs to understand what makes up the culture of the organization.<br />

The organization needs to identify what are the items or knowledge points that make the organization<br />

unique. It needs to create a process or method to understand the information customizations or<br />

specialties.<br />

The organization should document the use of standard KM practices such as lists, rules, expert<br />

systems, or story-telling?<br />

138


Mitchell Cochran<br />

Many organizations cannot run without a central control system. The organization needs to develop<br />

procedures to be able to operate in a limited form if the technology is not available. The understanding or<br />

knowledge of the processes will be critical for maintaining a manual methodology. An organization should<br />

always be able to issue payments with paper checks for a short duration. A longer duration might create<br />

problems for tracking expenses, sales, deliveries or other business tracking methodologies.<br />

4. Strategic knowledge<br />

The organization needs to be able to identify the various units of knowledge as part of the inventory. The<br />

knowledge or business processes need to be broken down into logical groups. The groups can be as<br />

large as business units or as small as any functional concern that cannot be recovered if lost. The key<br />

issue is the organization needs to identify what knowledge can be recreated and the cost to recover it.<br />

The concept is to identify the knowledge that is global in nature or easily available outside the<br />

organization. From a black box type of operation which consists of an input, process and output, the<br />

process is not necessarily known. The issue is for the organization to be able to determine when the<br />

process is known or unknown and its importance.<br />

It is just as important to identify the knowledge that is unimportant to the organization. Many business<br />

functions could be outsourced such as human resources, payroll, billing, or accounting. Many small<br />

businesses use a PC based accounting package such as QuickBooks for those tasks. The knowledge<br />

needs to be identified but it can easily be recreated if necessary. It is important to realize what knowledge<br />

can be recreated or does not need to be maintained. A messenger process may require a vehicle but the<br />

type of vehicle may be immaterial.<br />

The knowledge inventory is going to require that the organization determine which units are proprietary<br />

and how they are maintained. The resources need to be identified that are necessary to maintain that<br />

knowledge. Some knowledge units will be obtained from outside of the organization. The sources,<br />

processes, assumptions, and necessary resources need to be identified.<br />

5. Migration issues<br />

As the organization documents and understands its knowledge base, it will be better prepared for<br />

maintaining or migrating that knowledge to new applications. The organization will have created<br />

documentation that describes the inputs, processes, and desired outputs of each knowledge unit. It will<br />

understand the data relationships or models that have been used in the applications. An organization<br />

should consider the following activities to aid in migration of knowledge:<br />

Develop and maintain a complete data dictionary. The data dictionary needs to include knowledge<br />

that is created or maintained both internal and external to the organization.<br />

Use application tools which allow for the methodology to be visible. Application toolsets or power<br />

tools can provide power users with the capability to modify the code. Since the application logic is<br />

visible, it can be understood and migrated.<br />

The organization needs to document and understand the meta-data or assumptions that validate the<br />

knowledge.<br />

Have the outside vendor contractually obligated to create reports which can be used to export data to<br />

new applications.<br />

Have a vendor document what information will be lost in moving to a new application. Reports can be<br />

generated prior to migration to allow for the capturing of information that may be lost.<br />

A common migration issue is moving information from one system into another where the fields do not<br />

have the same configuration. There is the potential for information to be lost if the description of the new<br />

data element is smaller than the existing one. Consider a case where the last name of an individual is<br />

moved from a description of 25 alpha numeric characters to a field that is limited to 20 characters. There<br />

is the potential that a last name will be incorrect due to truncation of the data element after the migration<br />

process.<br />

The migration effect needs to evaluate if there are any non-standard implementations. In the case of the<br />

City of Monrovia, certain unused data fields were ‘borrowed’ to define custom City information. The<br />

custom information helps to create a unique service or resource for the organization that creates it.<br />

139


Mitchell Cochran<br />

However, as the application is updated or migrated the vendor will not be aware of how non-standard<br />

implementations might impact new levels of code.<br />

6. Conclusion<br />

Many organizations follow standardized business processes. In general, they understand how their<br />

business works and what their industry expertise is. The issue is if the organization documents or<br />

understands what makes the business unique. Many organizations don’t have a detailed inventory of the<br />

knowledge that is used by the operations of the business. Many authors have discussed the issues of<br />

tacit knowledge within the workforce. The issue described in this paper is determining what knowledge is<br />

tacit within the applications. Much of that tacit knowledge may not be available during disaster recovery<br />

or during a migration to a new application. The new vendor may or may not have the same knowledge<br />

base. The onus is on the customer to determine what knowledge will be lost and what will need to be<br />

migrated or created; caveat emptor or buyer beware.<br />

The organization needs to develop a knowledge inventory of the processes it uses. The knowledge base<br />

needs to be broken down into units that can be treated as business functions. The knowledge is identified<br />

as coming from either internal or external sources. The importance of the knowledge units also needs to<br />

be identified.<br />

The detailed knowledge inventory can be used as a base to understand what migration efforts will be<br />

needed when moving to new applications.<br />

References<br />

Badaracco, J.(1991) The knowledge link, Harvard Business School Press, p. 79.<br />

Bots, P.W.G. and H. de Bruiin, (2002) “Effective Knowledge Management in Professional Organizations: Going by<br />

the rules,” 35th Hawaii International <strong>Conference</strong> on System Sciences, IEEE Computer Society Press.<br />

Buhler, Patricia, (2002) “Building the Learning Organization for the 21st Century: A Necessary Challenge.<br />

SuperVision Vol. 63. Iss.12 pp. 20.<br />

Fuller, S. (2002) Knowledge management Foundations, Butterworth and Heinemann, Boston.<br />

Grossman, Martin, (2006) “An Overview of Knowledge Management Assessment Approaches”, The Journal of<br />

American Academy of Business, Vol. 8, Num. 2, March.<br />

KMEdge.org, (2010) APQC, “Overcoming Knowledge Loss: Key Knowledge Retention and Transfer Approaches”,<br />

Kmedge.org/wp/overcomingknowledge3.html.<br />

Leonard, Dorothy and Kiron, David, (2002) “Managing Knowledge and Learning at NASA and the Jet Propulsion<br />

Laboratory (JPL)”, Harvard Business Review, Sept 27.<br />

McManus, Denise J., Synder, Charles A., and Wilson, Larry T. (2003) “The Knowledge Management Imperative,”<br />

DSS 2004 <strong>Conference</strong> Proceedings,<br />

http://www.knowledgeharvesting.com/documents/KnowledgeManagementImperative.pdf,.<br />

Nelson, Richard R., Winter, Sidney G. (1982),An Evolutionary theory of Economic Change, Cambridge, MA: Belknap.<br />

Rastogi, P.N., (2000)” Knowledge management and intellectual capital – the new virtuous reality of competitiveness”,<br />

Human Systems Management, Vol, 19 No. 1. pp39-48, January 01.<br />

Sage, A.P. and W.B. Rouse, (1999) “Information Systems Frontiers in Knowledge Management,” Information<br />

Systems Frontiers, Volume 1, Number 3, pp. 205-219.<br />

140


Motivations for IT Outsourcing in Public Sector Local<br />

Government<br />

Michael Cox, Martyn Roberts and John Walton<br />

University of Portsmouth Business School, UK<br />

Michael.Cox@port.ac.uk<br />

Martyn.Roberts@port.ac.uk<br />

John.Walton@port.ac.uk<br />

Abstract: This paper examines the approach taken to Information Technology (IT) outsourcing in four local<br />

government councils in the UK. This is important because, whilst outsourcing has become a significant issue in the<br />

restructuring of organisations and is increasingly used within both the private and public sectors, there has been a<br />

lack of research into IT outsourcing in the public sector and particularly within local government. This paper provides<br />

an in-depth study into how outsourcing is managed in local councils and how successful it has been; especially<br />

considering its sometimes controversial nature and the mixed press results it receives. To complete this study,<br />

interviews, containing both qualitative and quantitative questions, were conducted with key people at the four<br />

councils. These interviews examined the rationale for IT outsourcing. The findings from the interviews were then<br />

compared to the current literature on IT outsourcing to identify best practice. This research shows that, whilst cost<br />

savings remain important, councils focus on achieving best value when outsourcing IT rather than simply lowest cost.<br />

Indeed, it shows that whilst outsourcing can result in improved efficiency, councils that focus primarily on cost<br />

savings are often less successful. However, whilst the results revealed that IT outsourcing was more successful at<br />

councils who focused on long-term strategic goals, the interviewees considered the strategic benefits of outsourcing<br />

less important than improving the service. The structured selection process that is imposed by legislation allows<br />

council managers to gain a better understanding of the outsourcing requirements and make informed decisions to<br />

achieve best value, however the need for cost efficiency can result in a more short-term focus. The cost of the<br />

process and its inflexibility makes it more difficult for councils to focus on long-term goals. The study concludes that,<br />

whilst councils recognise that both the contract and trust are important to ensure that outsourcing is successful, the<br />

culture of risk aversion in the public sector tends to lead to a ‘play it safe’ mentality resulting in an overemphasis on<br />

the contract. This can lead to a short-term focus that could make it difficult for the council and the provider to work<br />

together to meet long-term goals. The councils were generally skeptical of developing partnerships; however, the<br />

research reveals that councils who focused predominantly on the contract were less successful than those who<br />

developed partnerships with their providers. The authors therefore recommend that, in order to achieve greater<br />

success, councils should develop partnerships and focus on best value and long-term strategic goals when<br />

outsourcing IT.<br />

Keywords: Information Technology (IT); Information Systems (IS); outsourcing; public sector; local government.<br />

1. Introduction<br />

The use of outsourcing, although not a new phenomenon, has increased in recent years as firms seek to<br />

lower costs and increase efficiency in response to higher levels of global competition. Child (2005, p.179)<br />

describes outsourcing as “the contracting out of activities that need to be undertaken on a regular basis,<br />

which otherwise would be conducted within an organisation”. Kern and Willcocks (2000, p.322) define IT<br />

outsourcing as “a decision taken by an organization to contract out or sell the organizations IT assets,<br />

people and/or activities to a 3 rd party supplier, who in exchange provides and manages assets and<br />

services for monetary returns over an agreed time period”. Outsourcing has become a significant issue in<br />

the restructuring of organisations and many commentators agree that it is currently “one of the fastestgrowing”<br />

and most important activities in business (Burnes & Anastasiadis, 2003, p.355; Weinert &<br />

Meyer, 2005, p.1).<br />

Outsourcing offers numerous advantages; however there is also evidence that outsourcing often fails.<br />

The potential benefits include cost savings, efficiency gains, improved flexibility, access to world-class<br />

expertise and focus on core competencies. However, outsourcing also poses numerous risks that must<br />

be managed in order for outsourcing to be successful (Kremic, 2006, p.467).<br />

Outsourcing is identified in two categories; outsourcing of core value-chain operations and outsourcing of<br />

support activities. The core value-chain operation is the outsourcing of supply chain activities such as<br />

distribution, whereas outsourcing of support activities includes HRM and facilities management (Child,<br />

2005, p.181). Outsourcing initially involved the outsourcing of non-core activities to reduce costs and<br />

improve efficiency and had a relatively short-term focus. However, more recently outsourcing has been<br />

used more strategically as firms increasingly seek to become more efficient by focusing on areas where<br />

they can achieve competitive advantage (DiRomualdo & Gurbaxani, 1998, p.1). However, some theorists<br />

141


Michael Cox et al.<br />

suggest that you should not outsource your core competency, whilst others argue that it can be difficult to<br />

define what is core and non-core. Alexander and Young (1996), cited by Que’lin and Duhamel (2003,<br />

p.648), distinguish between activities that “are critical to performance” but only support the core and<br />

those that actually provide competitive advantage.<br />

This is a notable problem when outsourcing IT because, whilst some consider it a non-core activity, IT is<br />

becoming critical to service delivery. The fact that IT is complex and affects all activities within an<br />

organisation make it one of the most important activities outsourced (Weinert & Meyer, 2005, p.1). Lacity<br />

and Willcocks (1995, p.226) note that IT changes rapidly and switching costs to new technologies or<br />

suppliers can be high because IT demands high investment costs. They suggest this makes IT<br />

outsourcing unique compared to the outsourcing of other activities. This will have implications for how IT<br />

outsourcing is conducted and managed.<br />

1.1 Outsourcing in the public sector<br />

The public sector has followed the trend set by the private sector and IT outsourcing is now one of the<br />

most important outsourced activities that by “1996 had led to contracts worth £2 billion” (Lacity &<br />

Willcocks, 1997, p.85).<br />

In the 1980’s the government introduced Compulsory Competitive Tendering (CCT), this “involves<br />

government or firms using a competitive bidding process to help decide who should have the right to<br />

produce or deliver goods or services” within the public sector (Rimmer, 1994, p.79). This process was<br />

driven by “marketing and political trends” with the aim of making the public sector more competitive<br />

(Marco-Simó, et al, 2007, p.51). It was assumed that market-based competition would be able to provide<br />

improved efficiency and lower cost whereas public monopolies provided no incentives to improve<br />

performance (Besley & Ghatak, 2003). However, traditional outsourcing contracts in the public sector<br />

were often granted to the lowest priced bidder, which can lead to lower quality provision. Thus, the<br />

government updated CCT in 1998 to focus on best value rather than simply lowest cost, whether<br />

internally or externally (Butler, 2003).<br />

However, whilst outsourcing has become one of the most important tools in public sector management,<br />

failures tend to attract greater publicity, due to political accountability, whereas successes often go<br />

unnoticed (Maughan, 2003). These failures have led to an increasing interest in how outsourcing is<br />

managed in local councils and to what extent it has been successful.<br />

2. Literature review<br />

Much of the literature available deals with IT outsourcing in general without specifically addressing cases<br />

within the public sector or local councils. Indeed, Vilvovsky (2008, p.337) notes that the available<br />

research is “limited and fragmented”. De Looff (1996) finds this interesting since he shows that in the<br />

Netherlands 30% of IT outsourcing is accounted for by the public sector, a figure not too dissimilar to the<br />

UK. However, despite this, Marco-Simó et al (2007, p.52) note that outsourcing in the public sector “has<br />

not produced a level of research interest proportional to its economic importance”. This has led to an<br />

increased interest in how outsourcing is managed within the public sector; especially considering the<br />

increasing use of IT outsourcing within the public sector the mixed press results it receives.<br />

The lack of research in this area can be problematic for making comparisons. However, it does provide<br />

an opportunity to explore a gap in the literature and to discover if anything can be learnt from how<br />

outsourcing is managed in local councils. This is important because whilst the private and public sectors<br />

share some similarities they also have some notable differences. Vilvovsky (2008, p.338) notes that the<br />

public and private sectors are divided both “ideologically and operationally”. Both have similar values in<br />

terms of “responsiveness, honesty and accountability” but the public sector is not motivated primarily by<br />

financial performance. Thus, public sector managers may adopt a different mindset to outsourcing, based<br />

on goals and values other than cost efficiency, whilst being further constrained by political accountability.<br />

This could result in a more cautious approach to outsourcing based on minimising risk. Burnes and<br />

Anastasiadis (2003, p.365), in their comparison of a police force and an insurance company, further<br />

highlight differences in supplier selection, contractual arrangements and in the management of the<br />

relationship between the public and private sectors. One major difference is that public agencies are<br />

legally required to outsource any activity that can be done cheaper by outside contractors, as long as<br />

they meet the minimum specifications (Burnes & Anastasiadis, 2003, p.359). The private sector, in<br />

142


Michael Cox et al.<br />

contrast, has more choice over whether they outsource and to whom and can therefore do so more<br />

strategically and approach the best suppliers directly (Lin, et al, 2007, p.164).<br />

Currie (1996, p.226), in a comparison of the experiences of IT outsourcing in the UK’s private and public<br />

sectors, shows that IT outsourcing poses significant challenges to both and should not simply be seen as<br />

a “quick-fix” solution. Lacity and Willcocks (1997, p.106) investigated two case studies in the US public<br />

service. They have previously suggested that the UK has had “long standing problems” with outsourcing<br />

and that the results have often fallen short of expectations. In this article, they suggest that politicians<br />

often believe that outsourcing will automatically save money whereas in reality outsourcing either has<br />

failed or has yet to show any real benefits.<br />

Frederick (1994), however, contends that CCT has been successful and has produced both cost savings<br />

and increased efficiency. Indeed, Maughan (2003) suggests that the public sector is better at managing<br />

outsourcing relationships than the private sector because they devote staff exclusively to managing the<br />

relationship whilst the private sector generally does not which can create a lack of focus and clarity. He<br />

also states that the public sector benefit from having clear guidelines and a lengthy procurement process,<br />

which allows for multiple tenders, imposed upon them by legislation. Vilvovsky (2008, p.342) contends<br />

that public sector managers often find legislation and guidelines “a legal constraint rather than an aid”<br />

which may create inflexibility and lead to contracts being awarded to the lowest cost bidder rather than<br />

the best quality. Despite this, Burnes and Anastasiadis (2003, p.365) argue that the private sector could<br />

learn from the public sector’s structured approach. They suggest that government guidelines can create a<br />

greater understanding of outsourcing requirements and may produce more stable relationships. This<br />

research explores these issues further.<br />

3. Methodology<br />

This research takes the form of a multiple case study. Creswell defines a case study as “a single,<br />

bounded entity, studied in detail, with a variety of methods, over an extended period” (Maylor &<br />

Blackmon, 2005, p.243). Marco-Simó et al (2007, p.59) have argued that we need more “real cases”<br />

before “we can generalise the findings”. A case study approach was chosen because this type of<br />

research is useful when ‘how’ or ‘why’ questions are being asked (Saunders et al, 2007, p.139).<br />

The case study allows an in-depth examination to develop an enhanced understanding of the subject<br />

area and develops areas for further research (Flyvbjerg, 2004, p.420). Commentators agree that this<br />

method allows for a greater understanding of complex issues and is more comprehensive than surveys,<br />

due to its ability to “come closer….to the complexity of real organisation setting” (Maylor & Blackmon,<br />

2005, p.242).<br />

Yin (1984), cited by Gable (1994, p.113-115), believes that a single case study is useful in areas that are<br />

under-researched and require exploration, whereas multiple case studies are used to test for patterns<br />

and draw comparisons.<br />

The purpose of this research was to explore a complex issue in-depth and to identify some common<br />

trends rather than to present a generalised industry view. Indeed, Flyvbjerg (2004, p.420) argues that the<br />

case study is a reliable method in itself due to its ability to incorporate multiple data collection methods.<br />

3.1 Case studies<br />

The case studies are based on four local government councils in the south of the UK.<br />

Table 1: Description of the cases<br />

Council Description<br />

A Council A is a district council that serves a population of around 110,000 people. IT supports<br />

the council’s everyday operations and is critical to business performance and service delivery.<br />

IT is one of the most important activities outsourced at this council. A budget report meeting in<br />

2008 revealed that the contract for IT services is the second most expensive contract at the<br />

council, at over £700,000 per year. The ICT strategy at the council is not only to improve the<br />

efficiency and effectiveness of the service but also to increase its accessibility to all.<br />

The Council signed a 5-year contract with their service provider in 2005, worth £2.5 million.<br />

This contract was to provide the council with a full IT managed service that included desktops,<br />

servers, infrastructure and applications support.<br />

143


Michael Cox et al.<br />

Council Description<br />

B Council B is a unitary authority that serves a population of around 200,000 people. Council B<br />

chose to outsource to one supplier in a 10-year contract. This contract included not only the<br />

provision of IT but also related services such as HR, payroll, Customer Service, local taxation<br />

and benefits, procurement and property services. The council sought one strategic partner to<br />

ensure greater integration between the councils services but also because greater cost<br />

reductions could be achieved by outsourcing many services to the same provider. The main<br />

reason that the council opted for a long-term agreement was because of the high investment<br />

needed at the start of the agreement. This meant that the council needed to allow sufficient<br />

time for the provider to recoup their investment costs.<br />

C Council C is a unitary authority that serves a population of 150,000. This council outsourced for<br />

seven years. The council outsourced its entire IT department. Services provided include<br />

procurement, technology refresh, network development and support, offsite server hosting<br />

support and offsite mainframe services, as well as application support and management, 24x7<br />

services monitoring and helpdesk.<br />

D Council D is a district council that serves a population of 110,000. Council D do not outsource<br />

IT.<br />

3.2 Data collection methods<br />

To answer the research questions semi-structured interviews were conducted at four local councils. The<br />

interviews were held with the heads of IT at the councils. The interviews were the primary data collection<br />

method and allowed the authors to explore the interviewees’ experience and attitudes towards IT<br />

outsourcing. This paper presents a summary of the data collected.<br />

4. Analysis of results<br />

4.1 Motivations<br />

Table 2 summarizes the main issues as described by the four interviewees at the four councils.<br />

Table 2: Summary of responses<br />

Council A In 1998 Recruitment issues were the main reason to outsource because the council had difficulty<br />

attracting and retaining staff. However, this became less important after the 2005 outsourcing<br />

agreement began. Cost savings and improved quality were rated as the second and third most<br />

important reasons to outsource. However, outsourcing was also designed to set the strategic<br />

direction and improve the council’s responsiveness to the public.<br />

Council B The council primarily outsourced to improve quality and to gain much needed investment. Cost<br />

savings, however, were considered the second most important reason because the service had<br />

become costly to maintain. The council also outsourced to focus on strategy but this was<br />

considered the sixth most important reason.<br />

Council C Recruitment issues were the most important motivation to outsource IT, with improving quality and<br />

gaining access to expertise second and third. Cost savings were unimportant because the council<br />

had to initially pay more for the outsourced service to improve quality. The council also wanted<br />

outsourcing to support strategy, which is one reason they opted for a partnership agreement.<br />

Council D This council does not outsource IT. However, cost savings are considered the most important<br />

reason that councils outsource, but, in order of importance, cost savings, access to expertise,<br />

focus on core and improving quality are other important reasons. Also that councils outsource to<br />

pass on the responsibility and risk of managing the IT service.<br />

The interviewees from the four councils were asked to asked to rank the seven motivational criteria (7<br />

being the most important, 1 the least). Table 3 and Figure 1 illustrate the responses.<br />

144


Michael Cox et al.<br />

Table 3: Interviewees ranking of main motivations<br />

Interviewee A Interviewee B Interviewee C Interviewee D<br />

Aggregate<br />

Score<br />

External expertise 2 4 5 6 27<br />

Improved Service Quality 5 7 6 4 22<br />

Cost savings 6 6 2 7 21<br />

Cost Restructure 1 5 1 3 10<br />

Focus on core 4 2 4 5 15<br />

Increased flexibility 3 3 3 1 10<br />

Recruitment issues 7 1 7 2 17<br />

25<br />

20<br />

15<br />

10<br />

5<br />

0<br />

External<br />

Expertise<br />

The motivations for IT outsourcing<br />

Improved<br />

service<br />

quality<br />

Aggregate ranking<br />

Cost<br />

savings<br />

cost<br />

restructur<br />

e<br />

Focus on<br />

Core<br />

Flexibility Recruitme<br />

nt issues<br />

Aggregate ranking 17 2 2 2 1 10 15 10 17<br />

Figure 1: Summary of individual rankings<br />

In Table 4 and Figure 2 the interviewees from the four councils were asked to state whether the<br />

motivations were: Very Important, Important, Unimportant or Irrelevant to their decision to outsource.<br />

Table 4: Interviewees categorization of motivations<br />

Interviewee A Interviewee B Interviewee C Interviewee D<br />

External expertise Unimportant Important Important Important<br />

Improved Service<br />

Quality<br />

Important Very important Important Important<br />

Cost savings Important Very important Unimportant Very important<br />

Cost Restructure Irrelevant Important Irrelevant Unimportant<br />

Focus on core<br />

competencies<br />

Important Important Important Important<br />

Increased flexibility Important Important Unimportant Unimportant<br />

Recruitment issues Very important Unimportant Very important Unimportant<br />

145


Recruitment Issues<br />

Flexibility<br />

Focus on Core<br />

Cost Restructure<br />

Cost Savings<br />

Improved Service Quality<br />

External Expertise<br />

Michael Cox et al.<br />

The importance of the main motivational factors<br />

0% 50% 100%<br />

Figure 2: Importance of motivational factors<br />

4.2 Discussion<br />

4.2.1 Improved quality and cost reduction<br />

Very Important<br />

Important<br />

Unimportant<br />

Irrelevant<br />

Don’t Know<br />

This research shows that quality improvement and cost savings were the most important reasons to<br />

outsource IT at local councils (see Figure 1). Although cost savings were a major reason for outsourcing<br />

IT, improvements in quality of service often took priority meaning that cost savings could not always be<br />

achieved immediately.<br />

Several of the councils believed that they could have improved quality and reduced costs in-house given<br />

time, but that results can be achieved far quicker through outsourcing.<br />

Councils that focused excessively on price can result in the provider cutting back in other areas such as<br />

innovation and supporting the council strategically and have a negative long-term impact.<br />

Several councils sought to achieve best value. However, these cases showed that a focus on long-term<br />

goals and building a working relationship can lead to better results.<br />

Thus, whilst cost savings are important, councils should adopt a balanced approach to outsourcing to<br />

achieve best value rather than simply lowest cost. Indeed, this research shows that the results of<br />

outsourcing are not instantaneous and supports Daly’s conclusion (2001, p.13) that those that primarily<br />

focus on cost savings are often less successful than those who target “more value adding activities”.<br />

4.2.2 Access to expertise<br />

Access to external expertise was considered an important reason to outsource IT by three Interviewees<br />

and was overall considered the third most important reason. This is because IT now requires a range of<br />

different skills, not all of which councils can provide internally, and is constantly changing. In contrast,<br />

service providers specialise in IT and can therefore call upon a larger pool of skilled workers (Vilvovsky,<br />

2008, p.338).<br />

A number of interviewees commented that whilst gaining access to expertise is important, councils need<br />

to retain key skills in-house.<br />

146


4.2.3 Recruitment issues<br />

Michael Cox et al.<br />

Recruitment issues were the most important motivation for outsourcing in local government 10 years ago.<br />

This is evident at Council A and Council C, who both outsourced initially around 2000, who found that the<br />

inability to attract skilled staff, due to higher pay in the private sector, was the main reason to outsource.<br />

This was partly because the millennium bug had created a shortage in the supply of skilled IT staff and<br />

pushed wages up.<br />

Recruitment issues now appear to be less important. This could be because councils are beginning to<br />

recognise the importance of IT and the need to offer higher wages to attract skilled staff and have<br />

become more commercially orientated.<br />

4.2.4 Focus on core competence<br />

The councils also outsourced IT to transfer the responsibility and risk of managing routine activities so<br />

that the in-house team could spend more time and energy focusing on strategy. This allows councils to<br />

transfer responsibility for service failures, ongoing costs, and staffing whilst the council only have to<br />

manage the contract. All Interviewees considered this an important reason to outsource, although as<br />

Figure 2 shows none considered it the most important reason. This illustrates that improving the service<br />

by reducing costs and improving quality takes priority over strategic considerations at local councils.<br />

Indeed, both Interviewee A and B support Lacity and Hirschheim’s (1993, p.260) conclusion that IT is<br />

more complicated than other support functions that are traditionally outsourced, and therefore it takes<br />

longer than anticipated managing the contract and dealing with problems, thus resulting in less time to<br />

focus on strategy.<br />

The councils, however, are becoming increasingly reliant on IT and see IT as the key to transforming<br />

service delivery. Indeed, whilst some private companies may consider IT a support function, IT at local<br />

councils is critical to service delivery. The councils, thus, expected outsourcing to support strategy.<br />

Council B, for instance, signed a long-term Strategic Services Partnership (SSP) with their provider,<br />

which was designed to create greater integration between departments and to transform service delivery<br />

(Whitfield, 2008, p.3). Council C, similarly, wanted their provider to support them strategically. The<br />

councils that opted for long-term contracts concentrated more on improving quality and transforming the<br />

service. However, even Council A, who outsourced primarily to reduce costs, expected outsourcing to<br />

support strategy. Indeed, Interviewee A noted that the lack of strategic support was one of the main<br />

reasons outsourcing did not meet expectations.<br />

The use of outsourcing to support strategy could explain the recent increase in SSPs, signed between<br />

councils and service providers (Whitfield, 2008, p.4), and the increasing focus on shared services. This is<br />

where councils come together to pool resources, such as combined data centres, to achieve higher levels<br />

of efficiency in-house. This, combined with the recruitment of private sector managers with commercial<br />

experience, can give councils greater leverage when outsourcing and allow them to focus more on<br />

strategy.<br />

4.2.5 Flexibility and restructured costs<br />

The interviewees found that it was difficult to write flexible contracts and changes were subsequently<br />

difficult and costly to negotiate unless they were in the provider’s interest. One council reported improved<br />

flexibility. However, they only achieved flexibility after putting the contract aside and developing a<br />

partnership. This suggests that partnerships based on mutual benefit will make the provider more open to<br />

changes.<br />

Furthermore, whilst Gartner (2008) suggest that “changing the cost structure of IT” will allow firms to<br />

respond more quickly to changing needs, the councils did not consider cost restructuring an important<br />

reason to outsource IT. The interviewees acknowledged the benefits of paying only a monthly service<br />

charge. However, only one council outsourced to gain investment because they had suffered from<br />

numerous years of under-investment and needed investment to transform the service. This allowed the<br />

council to gain immediate investment but pay it off over the length of the agreement. This gives the<br />

provider greater incentive to improve the service although the council had to sign a long-term contract to<br />

enable the provider to recoup their investment costs.<br />

147


5. Conclusion<br />

Michael Cox et al.<br />

The literature review suggested that private companies are focusing less on cost reductions when<br />

outsourcing (DiRomualdo & Gurbaxani, 1998, p.3), whereas with local government cost savings remain<br />

very important to councils due to budget constraints. However, whilst Child (2005, p.187) argues that<br />

outsourcing offers “significant, often immediate cost savings”, this was not the case at two of the three<br />

councils that outsourced. This is because, as Daly (2001, p.12) notes, councils are now focusing on<br />

achieving best value and recognise that improving the quality of the service has an impact on price. This<br />

research supports Daly (2001, p.13) and Vilvovsky’s (2008, p.342) conclusion that those that focus<br />

primarily on cost savings are less successful than those who target more value adding activities.<br />

Although Hart et al (1997, p.1128) argue that cost reductions often take priority because quality is difficult<br />

to specify within contracts, two of the three councils that outsourced achieved greater success by<br />

balancing cost savings with improving quality to achieve best value (Daly, 2001, p.12).<br />

Indeed, whilst Lacity and Willcocks (1997, p.106) argue that outsourcing has often fallen short of<br />

expectations in the public sector, the three councils that outsourced all noted that outsourcing had<br />

improved quality and produced cost savings. This supports Frederick’s (1994) conclusion that CCT has<br />

been successful and has produced both cost savings and increased efficiency. However, whilst<br />

DiRomualdo and Gurbaxani (1998, p.3) argue that private sector organisations are focusing more on the<br />

strategic benefits of outsourcing, this research shows that improving quality and reducing costs are the<br />

most important reasons to outsource IT at local councils.<br />

The councils, did however, outsource to gain access to external expertise because, as Vilvovsky (2008,<br />

p.338) observes, IT has become increasingly complicated and councils do not have all the necessary<br />

skills in-house. The providers, in contrast, specialise in IT and can call upon a larger pool of skilled<br />

workers. However, the councils retained small in-house IT teams with key skills to allow them to develop<br />

new ideas and support strategy to reduce their dependence on their provider. Furthermore, recruitment<br />

issues are no longer the most important issues at local councils. Although Vilvovsky (2008, p.338) argues<br />

that the public sector pays lower wages, this research shows that councils are beginning to recognise the<br />

importance of IT and are becoming more commercially orientated and therefore willing to offer higher<br />

wages to attract the best workers.<br />

Finally, the research shows that significant time and effort needs to be taken to manage the service after<br />

outsourcing, and that outsourcing is seen very much as a way of making organizations responsive to<br />

business change.<br />

References<br />

Besley, T., & Ghatak, M. (2003). Incentives, choice and Accountability in the Provision of Public Services (the<br />

Institute for Fiscal Studies WP03/08). Retrieved April 8, 2009, from the Institute of Fiscal Studies website:<br />

http://www.ifs.org.uk/wps/wp0308.pdf.<br />

Burnes, B., & Anastasiadis, A. (2003). Outsourcing: a public-private sector comparison [Electronic version]. Supply<br />

Chain Management: An International Journal, 8 (4), 355–366.<br />

Butler, P. (2003). Timeline: outsourcing and the public sector. Retrieved April 17, 2009, from Guardian website:<br />

http://society.guardian.co.uk/microsite/outsourcing_/story/0,13230,933819,00.html.<br />

Child, J. (2005). Organization: contemporary principles and practice. Oxford: Blackwell.<br />

Currie, W. L. (1996). Outsourcing in the private and public sectors: an unpredictable IT strategy [Electronic version].<br />

<strong>European</strong> Journal of Information Systems, 4, 226–236.<br />

Daly, G. (2001). Outsourcing in Government – pathways to value, Retrieved July 26, 2009, from the Accenture<br />

website: http://www.accenture.com/NR/rdonlyres/438B5454-BEED-499F-A94E-<br />

2CFF389F2890/0/outsourcing.pdf.<br />

De Looff, L. A. (1996) IS outsourcing by Public Sector organisations. Paper presented at the International Federation<br />

for Information Processing World <strong>Conference</strong> on Advanced IT Tools, Australia. Retrieved April 2, 2009, from<br />

http://www.acs.org.au/president/1996/ifip96/i96iso.htm.<br />

DiRomualdo, A., & Gurbaxani, V. (1998). Strategic Intent for IT Outsourcing [Electronic version]. Sloan Management<br />

Review, 39 (4), 1-16.<br />

Flyvbjerg, B. (2004). Five misunderstandings about case-study research. In C. Seale, G. Gobo, J. Gubrium, & D.<br />

Silverman (eds.), Qualitative research practice (pp.420-434). London: SAGE.<br />

Frederick, D. (1994). Why Compulsory Competitive Tendering for Local Government Services is not as good as<br />

Privatisation. Retrieved April 6, 2009, from http://www.libertarian.co.uk/lapubs/econn/econn052.pdf.<br />

Gable, G. (1994). Integrating case study and survey research methods: an example in information systems<br />

[Electronic version]. <strong>European</strong> Journal of Information Systems, 3(2), 112-126..<br />

Gartner. (2008). Gartner says changing the cost structure of IT will become a business imperative for most CIOs.<br />

Retrieved June 15, 2009, from Gartner website: http://www.gartner.com/it/page.jsp?id=776812.<br />

148


Michael Cox et al.<br />

Hart, M. (2008). Is short-termism holding back public sector outsourcing? Retrieved August 19, 2009, from the Silicon<br />

website: http://www.silicon.com/publicsector/0,3800010403,39273472,00.htm..<br />

Kern, T., & Willcocks, L. (2000). Exploring information technology outsourcing relationships: theory and practice<br />

[Electronic version]. Journal of Strategic Information Systems, 9, 321-350.<br />

Kremic, T. (2006). Outsourcing decision support: a survey of benefits, risks, and decision factors [Electronic version].<br />

Supply Chain Management: An International Journal, 11 (6), 467–482.<br />

Lacity, M. C., & Hirschheim, R. (1993). Information Systems Outsourcing: Myths, Metaphors and Realities.<br />

Chichester: John Wiley & Sons Ltd.<br />

Lacity, M. C., & Willcocks, L. P. (1995). Interpreting Information Technology Sourcing Decisions From A Transaction<br />

Cost Perspective: Findings and Critique [Electronic version]. Accounting, Management and Information<br />

Technology, 5 (3/4), 203-244.<br />

Lacity, M.C., & Willcocks, L. (1997). Information systems sourcing: examining the privatization option in USA public<br />

administration [Electronic version]. Information Systems Journal, 7, 85–108.<br />

Lin, C., Pervan, G., & McDermid, D. (2007). Issues and recommendations in evaluating and managing the benefits of<br />

public sector IS/IT Outsourcing [Electronic version]. Information Technology & People, 20 (2), 161-183.<br />

Marco-Simó, J M., Macau-Nadal, R., & Pastor-Collado, J. A. (2007). Information Systems Outsourcing in Public<br />

Administration: An Emergent Research Topic. Paper presented at the <strong>European</strong> and Mediterranean<br />

<strong>Conference</strong> on Information Systems. (pp.51-61).<br />

Maughan, A. (2003). Stop laughing and start learning from public sector outsourcing. Retrieved April 8, 2009, from<br />

the Silicon website: http://comment.silicon.com/0,39024711,10003991,00.htm.<br />

Maylor, H., & Blackmon, K. (2005). Researching business and management. Basingstoke: Palgrave Macmillan.<br />

Que’Lin, B., & Duhamel, F. (2003). Bringing Together Strategic Outsourcing and Corporate Strategy: Outsourcing<br />

Motives and Risks [Electronic version]. <strong>European</strong> Management Journal, 21 (5), 647–661.<br />

Rimmer, S. J. (1994). Competitive Tendering and Contracting: Theory and Research [Electronic version]. The<br />

Australian Economic Review, 27 (3), 79-85.<br />

Saunders, M., Lewis, P., & Thornhill, A. (2007). Research methods for business students (4 th ed.). Harlow: Financial<br />

Times Prentice Hall.<br />

Vilvovsky, S. (2008). Differences between public and private IT outsourcing: common themes in the literature. Paper<br />

presented at the Proceedings of the 2008 international conference on Digital government research,<br />

International <strong>Conference</strong> Proceeding Series (337-346). Montreal, Canada.<br />

Weinert, S., & Meyer, K. (2005). The evolution of IT outsourcing: From its origins to current and future trends.<br />

Whitfield, D. (2008). PPP DATABASE: Strategic Service-Delivery Partnerships for local authority ICT, corporate and<br />

technical services in Britain.<br />

149


Documenting Innovation: A Methodological Proposal and<br />

Application<br />

Stefano De Falco<br />

University of Naples, Federico II, Italy<br />

sdefalco@unina.it<br />

Abstract: Nowadays the business is driven by a constant and ever-changing demand for innovation. In order to<br />

remain competitive the companies must know how to present new products, processes and services to respond<br />

more quickly and effectively to the needs of consumers. Today as never before, companies have faced a similar<br />

challenge, to hone the best of his ability to innovate. In this highly competitive business environment, only those with<br />

managers and employees highly motivated and prepared, and can count on a creative process properly, it can be<br />

said to be ready to conquer success. However, if this is the scenario in which to move in order to achieve the<br />

objectives of efficaciousness, managers also need to consider efficiency aspects. Often the actions and the projects<br />

designed to introduce elements of innovation within companies require a lot of energy. We just refer to this term<br />

borrowed from physics that makes the concept of power’s absorption for a certain time corresponding to the different<br />

phases of the innovation process with regard to human resources, construction, infrastructure, equipment and<br />

financial requirements involved in the innovation process. But this power’s consumption can be optimized and thus<br />

reduced, with the same results (or to remain in physical jargon) of charge to be met, through a system of<br />

documentation of the actions of innovation introduced over time by the company. In this paper a methodology for<br />

documenting innovation, as a strategic key factor for firms, is developed and practical application is provided through<br />

the contribution of a young company working on information systems, the “Solvendo Company”(www.solvendo.net)<br />

for the construction of an information system finalized to innovation’s documentation. It is also offered an application<br />

based on RFID, radio frequency devices, for the process of data acquisition instrumentation, machinery, plant, in<br />

order to have a quick reference tool for operating on a single component that aggregates the “query- innovation<br />

processes “ activity over time.<br />

Keywords: innovation, tracking, tracciability, rintracciability, Rfid<br />

1. Documenting Innovation: A technology overview<br />

The analysis of the competitive landscape in which companies move, has highlighted the increasingly<br />

urgent need, in this global economy, to trace the development of innovative activities from within the<br />

company to the end customer, helping to control the production and quality , attesting the originality of the<br />

product and its features, certifying that the product / service has been built following the existing rules on<br />

employment protection and health, and enabling real-time communication between businesses and<br />

business customers in partnership. This need can be answered only using the technology infrastructure<br />

in terms of all modern tracking and communications devices. (Top Management Forum on Strategic<br />

Management of Technology and Innovation 2010).<br />

This approach allows managers to start always from a step higher each time there is a new innovation<br />

process, avoiding to start from scratch with no memory of previous interventions.<br />

To frame the issue in the context of this analysis, in the following a summary of the opportunities that<br />

ICTs offer for SMEs to innovate their processes, is reported.<br />

The evolution and integration among the various ICT (Web, Mobile & Wireless, RFID, etc.) always has an<br />

impact more significant and pervasive within all main processes of the enterprise value chain, including<br />

the most complex and cross (Supply Chain Management, Customer Relationship Management,<br />

Knowledge Management, etc..) and provides new channels of communication with customers of the<br />

company as the possibility of communication via the Web, online marketing tools, the Mobile CRM<br />

services, etc.. that are profoundly changing the relationships between the company and its customers.<br />

But these benefits are exploited only if you have a system to document over time the improvements<br />

offered by the technology that determines them.<br />

The development of applications mentioned above, however, is conditioned by the necessity of<br />

adjustment of basic infrastructure:<br />

Infrastructure hardware<br />

This infrastructure consists of:<br />

150


Stefano De Falco<br />

Systems used to collect and manage information, access to applications and personal productivity,<br />

such as servers, clients, mobile devices, etc.<br />

The basic software and environments found in server systems, such as operating systems, software<br />

for database management (DBMS) and application server environments<br />

(For example, the Web server enables you to publish web applications);<br />

Peripheral devices (printers, scanners, plotters, etc. RFID transponders and readers.)<br />

Systems designed to ensure reliability of services and data management, enable the automatic<br />

replication of critical data on both the server and / or client (backup) or storage and sharing through<br />

the use of large volumes of network data.<br />

Security systems<br />

Even the SMEs in their computerization process, can not neglect the security aspect both provided by the<br />

instruments "centralized" (for the protection of servers and shared computing resources) and both for the<br />

protection of the resources used by individual users.<br />

Communications infrastructure<br />

Systems Voice over IP (VoIP) have become increasingly important due to a number of the benefits these<br />

systems can offer, in terms of reducing the costs of new advanced services enabled. The types of VoIP<br />

systems used by the company are essentially three types:<br />

VoIP on PC is the use of communications software (eg Skype, MS Messenger, etc..) Installed on<br />

each PC connected to the Internet. The installation is free and simple to use;<br />

Systems-based adapters, adapters are used, connected to the line data and analog phones to the<br />

company, making the conversion from analog to digital. The approach turns out to be so cheap for<br />

the company, which can preserve their traditional phones, but sometimes with poor quality<br />

communication;<br />

PBX systems based on IP-PBX, PBXs same mode allowing connectivity to mobile VoIP over the<br />

traditional to the new digital phones. This solution enables you to preserve the investments already<br />

incurred and to migrate in a flexible and gradual to VoIP technology.<br />

The benefits from the introduction of VoIP systems consist in the reduction of costs:<br />

Communication costs, because calls between IP-connected sites are free, while calls to the outside<br />

world does not have significant additional costs compared to the use of data network and there is<br />

also the possibility of using GSM dual mode mobile / Wi -fi within the walls of company headquarters,<br />

with a further reduction of the cost of mobile telephony;<br />

Infrastructure-management costs, thanks to a general simplification of systems and services and the<br />

use of a single infrastructure for voice and data and the ability to use advanced services such as:<br />

Video conferencing costs (at significantly lower costs than traditional systems);<br />

Messaging costs (the ability to manage e-mail address in an integrated service with voice mail) and<br />

the ability to share files and documents while communicating.<br />

Business applications can be distinguished:<br />

Management systems;<br />

Business intelligence applications;<br />

Management systems<br />

a) simple software package.<br />

Functionality is relatively limited and generalized (administration, warehouse, manufacturing). These<br />

packages can be integrated with significant customization is needed to extend the functional coverage in<br />

some areas of the company or to accompany the growth in size and complexity of the undertaking.<br />

b) vertical systems and / or developed ad hoc.<br />

Are software applications developed for the specific needs of the enterprise.<br />

151


Stefano De Falco<br />

These applications are usually run on proprietary architectures. The development is generally performed<br />

by software houses side by side, in the case of more structured companies, by the internal IT function.<br />

These systems offer a good level of support built in, but having been developed over time, are often<br />

based on outdated technology. Some software companies linked to specific industrial and regional<br />

contexts have also developed packages "vertical" that, based on a solid knowledge sector, it reduces the<br />

risk of adoption, in terms of time and budget, from customers.<br />

c) Systems with a high level of functional integration.<br />

They are made by software houses national (or domestic), which deal directly with end customers install<br />

or use a limited number of partners. Distinguishing factor of these applications and adherence to the<br />

specific regulatory, operational and business of the Italian context.<br />

At the most complete enterprise integration are the ERP systems offered by IT companies that typically<br />

operate on international markets (eg SAP) ERP systems that are capable of supporting the whole key<br />

business processes in an integrated manner.<br />

The introduction of an integrated management system pushes the company to analyze their processes,<br />

and this is the first step to identify opportunities for improvement. Only by learning to structure the<br />

activities and information flows associated with them, and then introducing operating logic of explicit<br />

processes, the company may lay the basis for identifying indicators of efficiency and effectiveness on<br />

which then act to increase their level of competitiveness to improve the performance parameters<br />

(reduction of delivery time, increase productivity in back-office activities (administration, order<br />

management, purchasing management, etc..), reduction of errors in direct activities (production,<br />

deliveries, sales, etc..) and indirect.<br />

Generally, the management system plays an important role in supporting the change in the company.<br />

From this it appears evident the importance of flexibility in the management system adopted. As more the<br />

system is prepared for the expansion of its functional coverage, the more the management system<br />

becomes an asset for the company to guarantee flexibility (adaptation to the logical management) in<br />

terms of opening opportunities for change offered by evolving business scenarios.<br />

Applications of Business Intelligence (BI)<br />

BI applications are intended to exploit the massive amounts of data (commercial transactions, financial<br />

and administrative, navigation patterns on web email, text and hypertext, etc..) available to the company.<br />

These applications are based theTechnologies DWH / ETL for the collection and storage should be off<br />

line of dates collected, provide access to the dates from different funds to analyze, structure, calculation<br />

of appropriate indicators and generate reports and dashboards with the aim of supporting better<br />

decisions by management enterprise.<br />

Examples of applications are:<br />

Analysis of the overall performance of the company: they allow the employer and / or top<br />

management to monitor key performance (as measured through appropriate KPIs) of the enterprise<br />

as a whole and the level of individual business units or corporate functions;<br />

Business analysis: to analyze the dates for customers, sales and marketing;<br />

Financial analysis, to control the flow of cash input / output, trade receivables and payables and any<br />

critical issues identified and anticipated;<br />

Analysis of suppliers to analyze the dates relating to the supply chain (procurement and inbound<br />

logistics), measuring the basic performance (number of suppliers, costs, delivery times etc..).<br />

The benefits of BI applications, therefore, consist essentially offering the user to quickly and appropriately<br />

to their needs decision-making information and data in large quantities but easily accessible, allowing you<br />

to monitor over time the outcomes of decisions. These features are not offered by "traditional" software<br />

because:<br />

They are often given the use of the information residing on different databases (eg, customer<br />

database, it is used by the CRM system, it's products in the database management, etc.). BI<br />

applications can simultaneously access data instead of these dates in different bases, to analyze and<br />

integrate information structure of decision support;<br />

152


Stefano De Falco<br />

The information presented in the various applications are typically structured in<br />

Analytical and synthetic form, the business intelligence applications, allow instead of synthesizing<br />

and presenting such information through an effective graphic form;<br />

Requires analysis of the relationship between different variables, the analysis is not possible with<br />

software applications management "traditional".<br />

CAD to PLM<br />

The CAD and Product Lifecycle Management (PLM) are used to support the activities of product design<br />

and, generally, created to manage the entire lifecycle of the product itself, from its birth until after the<br />

sale.<br />

Alongside the traditional CAD in recent years has significantly developed the methodology / 3D display<br />

technology, enables the realistic simulation of a product from different points of view (style, design,<br />

production and even maintenance), providing considerable support to different decisions on product<br />

development.<br />

In addition, CAD systems have enabled more and more communication with other systems Computer<br />

Aided (CAX) systems such as Computer Aided Styling (CAS) used by product designers and CAM<br />

systems for the automatic generation of programs in machine control numerically.<br />

The PLM is the integration of various ICT applications generate and use the product data from<br />

engineering systems (CAD described above), the software supporting the organization of their work<br />

(PDM - Product Data Management), the document management systems. Within the PLM approach, you<br />

can define different "features" on managing the lifecycle of the product: these features ranging from<br />

collaborative design and distributed to the management of the bill of product in its different views (the<br />

view of the designer For example, different from that of those who make technical assistance), to portfolio<br />

management and product configurations.<br />

CAD systems can reduce design time and improve the quality of the project through the possibility of a<br />

better view of the project, the automatic production of drawings, the easier management of revisions of<br />

the project. PDM systems and "tending to PLM" is usually taken with the specific objective-or to "put<br />

order" in the product data and manage the communication between different actors. In some cases, the<br />

PLM system will increase communication between the technical department and other business<br />

functions, or even with external actors (co-designers and component suppliers), while in others it may<br />

support a more thorough cost analysis life cycle of a product.<br />

Web applications and online services<br />

Web applications and online services typically provide features not provided by traditional management<br />

systems. The company may use their own web applications, developed and customized based on your<br />

needs (Intranet), or use services provided by external providers, are all available remotely as an ASP<br />

(application service provider).<br />

The intranet can enhance such efficiency and the management of the business records of sales activities.<br />

Of particular interest to SMEs are:<br />

The use of shared platform in ASP mode on the standard of services requiring low levels of<br />

personalization (such as management of pay slips);<br />

Services Customer Relationship Management (CRM) for managing customer relationships, which<br />

relate to specific activities such as building a customer database, the creation of reports and<br />

statistics, newsletters, etc..<br />

The use of ASP services online it has the added advantage compared to the intranet, you use these<br />

applications at a cost linked to the real use, freeing the company from start-up costs of investment and<br />

maintenance of software.<br />

Web applications and services online B2B<br />

It intends to intercompany B2B set processes supported by customer-supplier Web applications. We can<br />

then distinguish the applications and services based on the type of process supported.<br />

153


1) Buying (e-Sourcing, e-Catalog)<br />

Stefano De Falco<br />

The e-Sourcing applications are applications and services to find new suppliers, their qualification and<br />

certification. The services and applications to enable e-Catalog:<br />

Increased efficiency of the purchasing process, as a result of both automation of some activities,<br />

such as the handling of request tobuying, workflow and approval process of the cycle of<br />

administration and accounting, both of "Reversal" in a logical self-service, certain activities directly<br />

external user;<br />

A reduction in delivery times, thanks to a streamlining of the overall management process and the<br />

request to purchase order (thanks to the delegation of purchase) and a possible reduction in the time<br />

of receipt of goods in the case of special delivery arrangements with suppliers;<br />

A reduction of maintenance costs in stock.<br />

Also in the field of e-Sourcing and e-Catalog is willing to operate the offer its technology solutions in ASP<br />

mode with value-added services related to the use of the platform. The goal of these operators and to<br />

allow SME users to benefit from advanced technology without having to develop in-house.<br />

2) sales<br />

These applications enable the company to market the products to its business customers through the use<br />

of Web catalogs.<br />

Also in this case, the company may decide to develop their own application or extranet use the services<br />

provided by external providers. Extranet applications sales support, in general, the publication of the<br />

catalog of products, with an option for business customers (eg other users of the product on the market<br />

or the distribution network operators) to search for items of interest, to place orders online, to monitor the<br />

progress of the order and receive a set of pre-and post-sales.<br />

The main benefits arising from the use of catalogs are:<br />

Possibility to reduce non-value added activities involving, for example, inclusion of net sales orders<br />

from within the management system, etc.<br />

Reducing the risk of errors in the compilation of trade documents (Eg, orders, etc.).<br />

An increase in the level of service to their clients, with the possibility of offering support services to<br />

the selection and customization of products, services, order tracking and after sales services to<br />

support your use of the product.<br />

3) information exchange and collaboration in the supply chain<br />

These applications extranet or B2B services provided by specific operators, which support the exchange<br />

of information and / or documents between the company and its supply chain partners (eg, contractors,<br />

suppliers, etc.).<br />

Extranet or use the services provided by external providers that, in recent years, played a particularly<br />

important role. These operators provide, in particular SMEs in industrial districts, services that support the<br />

exchange of documents (eg, work orders, notes, business documents) and informations through the use<br />

of a shared platform, based on technology Internet.<br />

The main benefits are:<br />

An increase in labor productivity of people involved in operational interface is the front office to back<br />

office;<br />

Improve the quality of processes, through a reduction of errors in documents;<br />

A more timely exchange of information;<br />

Reduced costs of communication with their supply chain partners.<br />

Web applications and services online B2C<br />

154


Stefano De Falco<br />

The main applications or services that the firm can use to support the relationship with consumers and,<br />

more generally, with all those (stakeholders) that have some interest in the enterprise and that can be<br />

classified into three main categories:<br />

i) Institutional sites: they serve to provide a range of information to stakeholders on the company:<br />

slag, products / services, financial data, job opportunities, etc.. These applications are simple and<br />

inexpensive that have no particular impact on the processes and organization of the company and<br />

therefore are particularly popular even among SMEs;<br />

ii) Places of service: used primarily for the provision of services on line of pre-sales and / or post-sale;<br />

iii) e-commerce sites used in a "eCommerce Service Provider".<br />

Again the company intends to market its products through the internet to end consumers have the choice<br />

of a site e-business owner, as his shop "virtual", or the services offered by "eCommerce Service<br />

Provider", bringing together in one place a set of virtual stores, offering a range of common services<br />

(marketing, payment, delivery, etc...)<br />

The advantage for an SME to use the eCommerce Service Provider is the ability to leverage the brand,<br />

visibility and the traffic generated from them, the were able to develop a broad community of users. In<br />

particular, through this new commercial channel an SME may be able to:<br />

Extend their reach the market from traditional channels;<br />

It can better serve its customer base in a logic CRM, through the online channel by providing<br />

increased and more timely information (about products, news and promotions) as well as an<br />

additional way, convenient and always available to carry out orders;<br />

The possibility of increasing the effectiveness of distance selling, and the case of operators<br />

specialized in selling via catalog, telephone, television, etc.., The present, already in their traditional<br />

business model and a value chain particularly suitable for sale on the Web.<br />

The internet offers the opportunity to promote products and brands on the market particularly effective<br />

way. The main marketing tools for Internet-based are:<br />

Purchase of keywords: provides for the promotion of a website on engines research through the<br />

purchase of sponsored links appear in the results of research generated by visitors through the use<br />

of specific keywords. Sponsored links consist of a title, a brief description of the product c / o service<br />

and a link to the Punta advertised site. Paid according to the number of visitors actually click on the<br />

link, the ads are sorted based on what the advertiser is willing to pay for each click received;<br />

Banner, net consists publish graphic panels of various sizes form (banner), which can be positioned<br />

in different parts of a web page, textual or visual content, a fixed or rotating. Banners can also be<br />

inserted in the newsletter, the are sent to a database of members, usually focused on a specific<br />

theme.<br />

Direct eMailing, this tool is based on sending emails to potential (or regular) customers using its CRM<br />

or buying from specialized society, lists of email addresses of the people have expressed their<br />

consent to receive advertisements.<br />

Mobile and Wireless Applications (M & W)<br />

Have applications, taking a growing interest in the support business processes for SMEs. M & W<br />

applications are based on:<br />

Cellular network, whatever the term used (cell phone, smartphone, palmate, finished industrial<br />

notebook with a connect card, box or trasponder3);<br />

Red Wi-Fi, no matter what it terms used (smartphone, palmate, the Hindu-dustry, laptop, box or<br />

transponder);<br />

RFID technology.<br />

These technologies can be of support to:<br />

1) Activities to be carried on cameo as: sales (purchase orders via mobile devices (cell phones,<br />

PDAs, laptops), a collection of dates for operational marketing activities (merchandising), technical<br />

support and maintenance, etc.;<br />

155


Stefano De Falco<br />

2) operational activities aimed at monitoring progress (tracking) workmanship. Through terminals<br />

equipped with an optical reader, or RFID, placed in appropriate positions along the production line<br />

inside or at the contractors / suppliers / clients are automatically read the information contained tag /<br />

transponder (put on products or pallets) as, for example, the progress of the production process, the<br />

outcome of the tests performed by the workers, the information on the subsequent stages of the<br />

manufacturing process.<br />

This overview is a preliminary useful step in the proposed methodology to characterize the technology<br />

level adopted by the firm to document innovation.<br />

2. Track and trace innovation<br />

The first step in the application of a methodology aimed at using technology as a tool for documenting<br />

innovation the enterprise knowledge, is the management capability comprehensively from the two<br />

aspects of supporting capability from enterprise knowledge organization and technology environment and<br />

three aspects of knowledge production, diffusion and application in knowledge activities system by dint of<br />

method of knowledge audit, designed evaluating index system of knowledge management capability and<br />

evaluated the knowledge management capability by practical case. (Suo; Wang; Zhang 2009).<br />

The general objectives of the technology infrastructure are to obtain information about the flow of<br />

activities related to interventions of innovation in the enterprise.<br />

This means keeping track of the various activities performed on a single business process in terms of<br />

traceability from upstream to downstream, and traceability in the opposite direction.<br />

A innovation traceability system allows to relate the human resources and structural and infrastructure<br />

resources of a company with innovation activities developed over time, in terms of searches, research<br />

partners outside the company, calls for funding of programs research used for specific research activities,<br />

results (if any prototypes, patents and research agreements, conventions, etc.) (Sixto 2006).<br />

Moreover, this innovation traceability system, as a corollary, allows to define the responsibilities of units<br />

involved in each activity, among other things to improve quality control, to combat counterfeiting, to<br />

increase collaboration among units, to control the increasing complexity of the units and so on.<br />

To implement the proposed innovation traceability system, we consider each business unit as being<br />

composed of a set of processing sites and their operators. At each processing run specific activities<br />

(functions) on the parts compose the whole product or a product. The sites of processing may be related<br />

to the same company or different companies or business units. (Zhonghua and Aihua 2008).<br />

The traceability of innovation must be directed to those units.<br />

In a manufacturing site and the next may be an intermediate storage (referred to as "Buffer"), or a<br />

transport system (in case of processing stations belonging to different sites or different organizational<br />

units).<br />

In general, units can also be regarded as a group or batch of units, if the members of that group have the<br />

same production history. The innovation supply chain traceability is achieved by recording all the<br />

information about the research activities carried out on the unit, and allowing that information to flow<br />

throughout the chain. In this way, we reach the goal of tracking is the tracking unit.<br />

Track a unit is to identify the source going back in the chain, while drawing a unit is to follow the path in<br />

the chain from producer to consumer with regard to the effects of innovative research activities. Tracking<br />

can be also related to each portion of the product concerned from the research for its innovation, and as<br />

such may be capable of tracking each company has played a role in the formation of this portion and thus<br />

was in the innovation process.<br />

The traceability of innovation is therefore also a powerful instrument of control processes. The<br />

management of material flows in critical areas provides insight into a given time what processes are<br />

being developed in order to innovate, to identify the causes greater rapidity and safety of noncompliance,<br />

to manage operations in time and materials-there is the improvement and standardization of<br />

quality, cost reduction, the rationalization of logistics flows and processes. Thus, the traceability must not<br />

only allow the registration of providers, products and locations of materials contribute to the formation of<br />

156


Stefano De Falco<br />

the product, but also the association of other important data for process control, such as time, quantity,<br />

indicators, etc..<br />

Description of the technological infrastructure<br />

In the proposed approach we suggest the use of RFID technology to fast detect the past and present<br />

innovation activities implemented in the firm, as a key factor of success (Bradford and Florin 2003).<br />

RFID (Radio Frequency Identification) is considered as the main technology for realizing the ubiquitous<br />

environment by providing the identity to the physical object. (Rosenberg 2006). The RFID tag, which is<br />

consisted of an antenna, transceiver, and integrated circuit with memory, has been widely used for supply<br />

chain and inventory management to increase the business efficiency. Also, it can be used for ID badges,<br />

access control, fleet maintenance, equipment and parts maintenance, parking log access control, toll<br />

gate system, car tracking, manufacturing line control (RFID Implementation in Retail and Manufacturing<br />

IEEE RFID 2008), e-government services (Carter and Bélanger 2005) etc. The tag is attached or<br />

integrated to an object such as a product can be read by RFID reader and then we can identify the object<br />

with the information in the tag.<br />

In this approach this technology is used to give a tool for managers to know fast the existing links<br />

between firm’s resources and research processes and research partners over time (Figure 1) (Brown,<br />

and Russell 2007), (Gentag 2006).<br />

Figure 1: RFID application for tracking research processes<br />

In terms of infrastructure, the traceability system as a model for the documentation of business innovation<br />

is made up of different terminal unit (TU) equipped with a reader / writer and RFID tags, the acquired data<br />

and transmit them to a pin or storage unit (SU). The SU record data sent by UT. Processing unit (PU) is<br />

responsible to collect and analyze data from the SM. Although the data are distributed in different SU, the<br />

PU may ask for a SU only, which will contact the other SU on behalf of the PU. The TU can be mobile<br />

(such as palmate, with screen and keypad) or "bench" (only the reader), and connected to a SU wireless<br />

technology or cable network, respectively. The TU enable operators to collect and send data on the<br />

production flow. The innovation documentation platform offers the opportunity to pose several queries to<br />

the system, including use of keywords, to search specific innovation processes under development or<br />

completed and archived.<br />

This platform was developed in research partnership with the “Solvendo Company”. Solvendo is an italian<br />

software house whose main mission is finding and developing smart solutions to real problems of SME.<br />

The main activities concern document management and dematerialisation, ontology-based knowledge<br />

157


Stefano De Falco<br />

management and e-learning, business software, web applications development. Solvendo is actually<br />

involved in technology transfer activities with Federico II University of Naples and is committed in<br />

research projects.<br />

The operation of the platform is as follows: When a request comes of an innovation, both process or<br />

product or system innovation, following a call for research funding, or an agreement with a research<br />

institution, or a customer need, then every element involved in the modification, extension, or integration<br />

process, is identified and associated with that specific request of innovation and it is also associated with<br />

keywords to the action implemented. This builds a database of innovation characterized by many fields<br />

useful for next query (Symbol_Reader, Handheld RFID Readers, (2007). Thus to new each request of<br />

innovation is first carried out a query on the database of innovation to see if there were other actions can<br />

serve as a starting point for the current action and especially if there are already skilled human resources<br />

in the company on that precise issue and physical resources more already evolved into the current target.<br />

Also another big plus of the platform is the possibility to classify, over time, all successful and not<br />

successful innovation activities. Therefore, every new innovation activities on the question of innovation<br />

database can be done by keyword, by date, research partners, by choice of process and by result<br />

(positive or negative). In figure 2 a resource registration form is reported and in figure 3 a palmate reader<br />

used in the implementation of the proposed system is reported (EPC 2007 and EPC GLOBAL 2005).<br />

Figure 2: Resource registration form in the proposed tracking innovation platform<br />

Figure 3: Reader in the proposed tracking innovation platform<br />

158


3. An application<br />

Stefano De Falco<br />

The methodology was actually applied at an Italian medium / large firm that produces and distributes<br />

dairy products in Italy and abroad, named ICCA spa. With the helpful cooperation of the processes chief,<br />

the processes of the company have all been surveyed, have been traced all human resources and<br />

equipment of the industrial plant and were made the report, in the time window 2000-2010, all research<br />

projects involving the Icca company and their involvement with these resources, as reported in table I.<br />

Table 1: Innovation grid for “Icca” company<br />

Structural Infrastructural Material Human Existing Innovation Research<br />

Processes Processes Resources Resource Technology Process ID Partner (s)<br />

Preparation -Power supply. -Oven. Unit 1 VoIP R72010 Dep.of<br />

products<br />

-Fuel. -Hydraulic<br />

PLM<br />

Informatics<br />

-Cleaning of the press.<br />

University<br />

workplace.<br />

of Naples<br />

-Calibration<br />

equipments.<br />

-Maintenance<br />

equipments.<br />

-Communications<br />

-Quality Control.<br />

Federico II<br />

Packaging Power supply. Packaging Unit 2 VoIP G32009<br />

-Fuel.<br />

-Cleaning of the<br />

workplace.<br />

-Calibration<br />

equipments.<br />

-Maintenance<br />

equipments.<br />

-Communications<br />

-Quality Control.<br />

-Communications<br />

-Quality Control.<br />

equipment.<br />

CAD<br />

Distribution Product Delivery Vehicles Unit 3 CRM<br />

ERP<br />

-<br />

Sale In outsourcing In<br />

In<br />

- -<br />

outsourcing outsourcing<br />

Promotion In outsourcing In<br />

In<br />

- -<br />

outsourcing outsourcing<br />

4. Results and discussions<br />

A result of applying the methodology of innovation relates to a request by the network of distributors of<br />

the products of “Icca” Company. The request was the possibility of implementing a new telephony<br />

software for order management. The application of the methodology has allowed a quick feasibility study<br />

of the needs required by the network of distributors who provided a positive solution to the case, as the<br />

preceding process of innovation (in the database identified innovation as a process of innovation<br />

R72010) introduced the transition to voice over IP telephony system that facilitates the ability to<br />

implement the new software to manage orders. In order to have committed human resources already<br />

formed in the previous process of innovation and also contacted the Dep. of Informatics of University of<br />

Naples Federico II, which had provided advice on the previous trial.<br />

5. Conclusions<br />

The proposed work concerns the very topical theme of innovation in operational terms, and introduces a<br />

methodology with a case of a real application through development of a platform that can really make an<br />

improvement both for the users of 'innovation, ie companies, but also for the suppliers of upstream<br />

research for innovation, universities and research centers.<br />

In particular, the proposed work addresses the problem from an innovation point of view circular, in the<br />

sense that often sees the technology as a result of the innovation process, but is itself an instrument to<br />

facilitate it. In the approach proposed has chosen to use technology to acquire data using RFID devices<br />

with the aim of documenting all activities relating to innovation developed over time by the company.<br />

159


Stefano De Falco<br />

The pursuit of this objective requires a series of preliminary steps, principally, according to Porter's<br />

approach, the classification of all processes which is characterized by the company, both structural and<br />

infrastructural. This arrangement allows the operator of innovation to have a functional association<br />

between processes and resources affected by human activity and structural innovation.<br />

The advantages of a system for documenting innovation in the company consist of the opportunity for<br />

managers of innovation in the enterprise to optimize the choice of the most appropriate resources to<br />

specific objectives related to project financing facility agreements with research centers on innovation<br />

needs of their customers, since they already know the "state of innovation" of each element and also<br />

what process it is inserted.<br />

Acknowledgements<br />

The author thanks to Eng. Fabio Di Marino of “Solvendo” society, for his valuable contribution to this<br />

research work, and thanks to Icca Company for having granted to work on their business processes and<br />

systems to test the proposed methodology.<br />

References<br />

Asian Productivity Organization, Top Management Forum on Strategic Management of Technology and Innovation,<br />

Strategic Management of Technology and Innovation ©APO 2007, ISBN: 92-833-7063-5 Report of the APO<br />

January 2010.<br />

Bradford, M., and Florin, J. Examining the Role of Innovation Diffusion Factors on the Implementation Success of<br />

Enterprise Resource Planning Systems” International Journal of Accounting Information Systems, 4, (2003),<br />

pp.205-225.<br />

Brown, I., and J. Russell, Radio Frequency Identification Technology: An Exploratory Study on Adoption in the South<br />

African Retail Sector, International Journal of Information Management, 27,(2007), pp. 250–265<br />

Carter, L. and Bélanger, F., The utilization of e-government services: citizen trust, innovation and acceptance factors,<br />

Information Systems Journal, 15(1), (2005), pp. 5-25.<br />

EPC2007, EPC Information Services (EPCIS) Version 1.0 Specification, April, (2007).<br />

EPCglobal, The EPCglobal Architecture Framework, July (2005).<br />

Gentag_NFC, Gentag to Piggyback on NFC Technology to Read RFID Sensors, September, (2006).<br />

O. J. Sixto, Is near-field communication close to success?, IEEE Computer, vol. 39, p. 3, March (2006).<br />

RFID Implementation in Retail and Manufacturing, IEEE International <strong>Conference</strong> on RFID (IEEE RFID 2008), Las<br />

Vegas, April 16-17, (2008).<br />

B. Rosenberg, RFID : applications, security, and privacy. Upper Saddle River, NJ: Addison-Wesley, (2006).<br />

Suo Bai-min; Wang Jia-bin; Zhang Han-bing, Primarily Research on Knowledge Audit for Evaluating Enterprise<br />

Knowledge Management Capability Management and Service Science, 2009. MASS '09. International<br />

<strong>Conference</strong> on Digital Object Identifier: 10.1109/ICMSS.2009.5302226 Publication Year: 2009 , Page(s): 1 - 5<br />

Symbol_Reader, Handheld RFID Readers, (2007).<br />

Zhonghua Liao and Aihua Yang, Knowledge Sharing Based on Project Management, China Human Resources<br />

Development, vol. 222, 2008, pp. 95–98.<br />

160


Legal, Privacy, Security, Access and Regulatory Issues in<br />

Cloud Computing<br />

Nomusa Dlodlo<br />

CSIR – Meraka Institute, Pretoria, South Africa<br />

ndlodlo@csir.co.za<br />

Abstract: Cloud computing is a sufficiently new research area. Since it is in its development stages, the information<br />

on the field is slowly being compiled by the researchers and practitioners from their experiences. Some of the areas<br />

in which there is still a gap on reporting on legal, privacy, security, access and regulatory issues. This paper raises<br />

an awareness of legal, privacy, security, access and regulatory issues that are associated with the advent of cloud<br />

computing. An in-depth literature survey is conducted on these and an analysis is drawn from the issues that are<br />

identified through the literature survey. Recommendations are then given on how the issues identified in the analysis<br />

can be mitigated. These recommendations centre around the issues of policy interventions, standards, privacy and<br />

data protection, traffic and congestion management, business continuity planning, security and regulation. This<br />

research is an advancement of knowledge in that field and is meant to initiate further debate on cloud computing<br />

Keywords: cloud computing, security, privacy, regulation, legal issues, interoperability, business continuity planning<br />

1. Introduction<br />

The emergence of very large specialised data centres that host thousands of servers has created a<br />

surplus of computing resources that has come to be called the cloud. The cloud is the term for networked<br />

computers that distribute processing power, applications and large systems among machines. This<br />

means that, computing is no longer on local computers but on centralised facilities operated by third party<br />

compute and storage facilities (Foster, 2010). Cloud computing transforms once-expensive resources like<br />

disk storage and processing cycles into a readily-available cheap commodity. By deploying Information<br />

Technology (IT) infrastructure and services over the network, any organisation can purchase these<br />

resources on an as-needed basis and avoid capital costs of software and hardware. By offering<br />

enterprises the opportunity to decouple their IT needs and their infrastructure, cloud computing has the<br />

likely ability to offer enterprises long-term IT savings, including reducing infrastructure costs and offering<br />

pay-for-service models.<br />

One of the definitions of cloud computing given in literature is as follows (Cloud computing, 2009):<br />

“Cloud computing refers to both the applications delivered as services over the Internet and<br />

the hardware systems and software in the datacentres that provide those services. The<br />

services themselves have long been referred to as Software as a Service (SaaS). The<br />

datacentre hardware and software is what we will call the cloud. When a cloud is made<br />

available in a pay-as-you-go manner to the public, we call it a public cloud; the service being<br />

sold is utility computing. We use the term private cloud to refer to internal datacentres of a<br />

business or other organisations that are not made available to the public. Thus cloud<br />

computing is the sum of SaaS and utility computing, but does not normally include private<br />

clouds”<br />

The National Institute of Standards and Testing (NIST) defines cloud computing under 5 identified<br />

characteristics as follows (How cloud computing, 2010):<br />

On-demand self service, which allows business units to get the computing resources they need<br />

without having to go through the IT department<br />

Broad network access, which allows applications to be built in ways that align with how businesses<br />

operate today – mobile, multi-device, etc.<br />

Resource pooling, which allows for pooling of computing resources to serve multiple consumers<br />

Rapid elasticity, which allows for quick scalability or downsizing of resources depending on demand<br />

Measured service, which means that business units only pay for the computational resources they<br />

use. IT costs match business success<br />

Kushida, et.al (Kushida, 2010 ) give the operating definition of cloud computing as:<br />

“Cloud computing provides on-demand network access to a computing environment and<br />

computing resources delivered as services. There is elasticity in the resource provision for<br />

161


Nomusa Dlodlo<br />

users, which is allocated dynamically within providers’ datacentres. Payment schemes are<br />

typically pay-as-you-go models”.<br />

With the advent of the cloud arises legal issues and those of privacy, security, access and its regulation.<br />

This paper gives a review of these issues and how they can be mitigated.<br />

2. Problem statement<br />

Since cloud computing is a field that is in its development stages, the information on the field is slowly<br />

being compiled by the researchers and practitioners from their experiences. Some of the areas in which<br />

there is still a gap on reporting are on are legal , privacy, security, access and regulatory issues. This<br />

paper raises an awareness of legal, privacy, security, access and regulatory issues that are associated<br />

with the advent of cloud computing.<br />

2.1 Process<br />

An in-depth literature survey is conducted on the legal, privacy, security, access and regulatory issues of<br />

cloud computing and an analysis is drawn from the issues that are identified through the literature survey.<br />

Literature that is relevant to this article was compiled through Internet searches, and searches of<br />

databases of online journals and conference proceedings. The search was done randomly, on the basis<br />

of the associated keywords identified. Recommendations as identified from current literature are then<br />

given on how the issues identified in the analysis can be mitigated. These recommendations centre<br />

around the policy interventions, standards, privacy and data protection, traffic and congestion<br />

management, business continuity planning, security and regulation.<br />

2.2 Question and objectives<br />

The main research question in this paper is, “ What is the current state of affairs on the soft issues of<br />

cloud computing and what does the literature say on the recommended way forward to further the issues<br />

of cloud computing.”<br />

The objectives are:<br />

Identify the legal, privacy, security, access and regulatory issues prevailing currently in the area of<br />

cloud computing<br />

Identify recommendations on the way forward on these issues<br />

3. Legal implications of cloud computing<br />

(Legal implications, 2010; Legal issues, 2010; Wisdom of clouds, 2008)<br />

There is the issue of “reasonable security” in the cloud computing context, and potential liability arising<br />

out of security breaches in the cloud. A company that provides a service to handle the personal<br />

information of another organisation has the responsibility to ensure that there is reasonable security to<br />

protect personal and confidential information.<br />

The data centres of cloud service providers are located in various locations all over the world. That<br />

means data on the cloud could be stored in any country. The ‘physical location’ raises the question of<br />

legal governance over the data. In case of a conflict between the cloud vendor and the customer the<br />

question of which country’s court system will settle the dispute comes to the fore. In cases where there is<br />

a litigation, an organisation will have to deal with a third party cloud provider to gain access to information<br />

relevant to the litigation. Considering the multiple copies of data that may be created, stored, recompiled,<br />

reused, dispersed and reassembled, what constitutes a “record” for evidence may be difficult to grapple<br />

with the cloud.<br />

The number of trademark filings covering cloud computing brands, goods and services is increasing as<br />

companies seek to better position themselves for cloud computing branding and marketing efforts.<br />

Therefore ensuring the uniqueness of a trademark with the advent of a cloud has been further<br />

complicated.<br />

Sharing and transferring data within the cloud is a problem. Organisations are legally prohibited from<br />

transferring personal information to countries that do not provide the same level of protection with respect<br />

to personal information. That means cloud providers will not be in a position to make any contractual<br />

162


Nomusa Dlodlo<br />

promises to their clients because in many cases they cannot say which countries data will be transferred<br />

to and from.<br />

Systems are vulnerable to damage or interruption from earthquakes, terrorist attacks, flood, fires, etc.<br />

Customers have to ensure therefore that they are insured against loss of business due to such potential<br />

losses. This is essential. If there is a breach of privacy due to the fault of the cloud vendor, the carrier<br />

should be liable for the compensation. Vendors on the other hand should do their level best to meet<br />

service level targets committed with any customer.<br />

Ideally, the data that is of the customer’s creation is protected under the intellectual property rights of a<br />

country. Therefore there should be compensation for infringement. Customers own the data. No vendor<br />

can claim ownership of any data that is uploaded or associated with intellectual property. Customer data<br />

includes all data maintained by the customer<br />

In one frequently cited scenario (Gurav, 2010), a government agency presents a subpoena or search<br />

warrant to the third party that has possession of a customer’s data. Had they retained physical custody,<br />

the customer might still have been compelled to surrender the information, but at least would have been<br />

able to decide for themselves whether or not to contest the order. The third party service is presumably<br />

less likely to go to court on behalf of a customer. In some circumstances the customer might not even be<br />

informed that their documents have been released.<br />

With the cloud there is lesser privacy protection under the law. To search a house or office (including<br />

documents stored on a computer), police need a warrant of arrest. To get the information that is stored<br />

on a third party’s web server they only need a subpoena, which is easier to obtain. This kind of search<br />

can happen without the cloud customer’s knowledge.<br />

4. Security issues in cloud computing (hidden risks, 2010)<br />

Due to its distributed nature, the cloud results in weak security systems that are easy to break into. The<br />

security of the system is only as strong as the weakest user’s set-up. Weak password recovery<br />

workflows, phishing attacks, and keyloggers present bigger security risks. In collaborative web<br />

applications that are built for groups, like Google Apps or any web-based project management software,<br />

any breach of security spreads across all participants.<br />

In cloud computing an organisation’s data is locked-in and the third party in control. When you participate<br />

in the cloud, you depend on a third party to make decisions about your data and platforms. Cloud<br />

computing also comes with chances of server unavailability and account lock-out. When the Internet goes<br />

down, access to one’s data is cut off. An important measure of security often overlooked by companies is<br />

how much downtime a cloud service provider experiences. The client should request the provider’s<br />

reliability reports to determine whether these meet the requirements of their business. Exception<br />

monitoring systems is another important area which companies should ask their service providers about<br />

(Binning, 2010).<br />

The biggest concern with cloud computing is that it puts all of a company’s data and applications in one<br />

place. Businesses should be wary of putting sensitive company information in public clouds. They should<br />

instead stick to low-risk, low volume applications and build internal and private clouds to enable<br />

collaboration within the organisation and externally with partners.<br />

Security is one of the most often-cited objections to cloud computing (Zhang, 2010). Cloud users face<br />

security threats from both outside and inside the cloud. This responsibility is divided among the cloud<br />

user, the cloud vendor and any third party vendors that users rely on for security-sensitive software or<br />

configurations. The cloud user is responsible for application-level security. The cloud provider is<br />

responsible for physical security, and likely for enforcing external firewall policies. Security for<br />

intermediate layers of the software stack is shared between the user and the operator. Cloud providers<br />

must guard against theft or denial-of-service attacks by users. One last security concern is to protect the<br />

cloud user against the provider. The provider will by definition control the ‘bottom layer’ of the software<br />

stack, which effectively circumvents most known security techniques. Users also need to be protected<br />

from one another. The primary security mechanism in today’s clouds is virtualisation. It is a powerful<br />

defence, and protects against most attempts by users to attack one another or the underlying cloud<br />

infrastructure.<br />

163


Nomusa Dlodlo<br />

Virtualisation is the enabler technology for the cloud and uses physical resources such as a server which<br />

is divided into virtual resources called virtual machines (VM). Customers cannot protect their VMs on their<br />

own. Cloud service providers are making substantial effort to secure their systems, in order to minimise<br />

the threat of insider attacks and reinforce the confidence of customers. For example, they restrict access<br />

to hardware facilities, adopt stringent accountability and auditing procedures, and minimise the number of<br />

staff who have access to critical components of the infrastructure. Security management of virtualisation<br />

technologies is required to reduce the risk of security exposures and enable security policy enforcement.<br />

A cloud-specific security issue is that of running arbitrary VM images. This is only one aspect of making<br />

sure the right data is available to the right user at the right time. The Cloud Security Alliance is directly<br />

pursuing in addition, issues of authentication, authorisation, privacy, integrity and non-repudiation and<br />

data reliability and availability.<br />

According to Chow et.al. (Chow, 2009), security concerns are categorised as:<br />

Traditional security<br />

Availability<br />

Third party data control<br />

Traditional security concerns involve computer and network intrusions or attacks that will be made easier<br />

or possible by moving to the cloud (Chow, 2009). VM-level attacks are a problem because of potential<br />

vulnerabilities in the VM technology. There are also cloud provider vulnerabilities. These could be<br />

platform level such as SQL-injection or cross-site scripting. Phishers and other social engineers have a<br />

new attack vector. The cloud user must protect the infrastructure used to connect and interact with the<br />

cloud, a task complicated by the cloud being outside the firewall in many cases. The enterprise<br />

authentication and authorisation framework does not naturally extend into the cloud. Traditional digital<br />

forensic methodologies permit investigators to seize equipment and perform detailed analysis on the<br />

media and data recovered. The likelihood therefore of the data being removed, overwritten, deleted or<br />

destroyed by the perpetrator in this case is low.<br />

Availability concerns centre on critical applications and data being available (Chow, 2009). As with<br />

traditional security concerns, cloud providers argue that their server uptime compares well with the<br />

availability of the cloud user’s own data centres. There are more single points of failure and attacks in the<br />

cloud. They may lack an assurance of computational integrity.<br />

The legal implications of data and applications being held by a third party are complex (Chow, 2009).<br />

Therefore there are many questions that remain unanswered. There is also a potential lack of control and<br />

transparency when a third party holds the data. If served a subpoena can a cloud user compel the cloud<br />

provider to respond in the required time frame? How can a cloud user be guaranteed that data has been<br />

deleted by the cloud provider? Audit difficulty is another side effect of the lack of control in the cloud. Is<br />

there sufficient transparency in the operations of the cloud provider for auditing purposes? There are<br />

contractual obligation issues also. One problem with using another company’s infrastructure besides the<br />

uncertain alignment of interests is that there may be surprising legal implications. Cloud provider<br />

espionage is the worry of the theft of company proprietary information by the cloud provider. How can a<br />

cloud user avoid data lock-in? The data itself might locked in proprietary format and there are also issues<br />

with training and processes. There is also a problem of the cloud user having no control over frequent<br />

changes in cloud-based services. Another possible concern is that the contracted cloud provider might<br />

use subcontractors, over whom the cloud user has even less control and who must also be trusted.<br />

5. Data access and interoperability<br />

The issue of data access and interoperability continues to be an outstanding matter for inherently<br />

distributed applications and federated organisations. Common best practices and standards are needed<br />

to achieve the fundamental properties of portability and interoperability for cloud applications and<br />

environments.<br />

A major challenge of moving applications to the cloud for most organisations is the need to master<br />

multiple languages and operating environments (Gurav, 2010). In many cloud applications a back-end<br />

process relies on a relational database, so part of the code is written in SQL, or other query language. On<br />

the client side, program logic is likely to be implemented in JavaScript embedded within HTML<br />

documents. Standing between the database and the client is a server application that might be written in<br />

a scripting language (such as PHP, Java and Python). Information exchanged between the various layers<br />

164


Nomusa Dlodlo<br />

is likely to be encoded in some variation of XML. Any web application needs to be available to legitimate<br />

visitors from all over the world. A true cloud spans the entire globe, with a server presence in multiple<br />

simultaneous locations.<br />

Besides technical issues, a cloud provider could suffer outages for non-technical reasons, including going<br />

out of business or being the target of regulatory action. Therefore organisations should be wary of this<br />

and put in place measures to ensure business continuity and service availability when outages occur<br />

(Zhang, 2010).<br />

Data lock-in by the service provider is a contentious issue for the customer. Software stacks have<br />

improved interoperability among platforms, but the storage Application programming Interfaces (APIs) for<br />

cloud computing are still proprietary. Thus, customers cannot extract their data and programs from one<br />

site to run on another. This has prevented some organisations from adopting cloud computing. Customer<br />

lock-in may be attractive to cloud computing providers, but their users are vulnerable to price increases,<br />

to reliability problems, or even to providers going out of business (Zhang, 2010).<br />

Applications continue to be more data intensive. Data bottlenecks are likely to occur as more users<br />

subscribe to a cloud service. Cloud users and cloud providers have to think about the implications of<br />

placement and traffic at every level of the system if they want to minimise costs (Zhang, 2010).<br />

6. Privacy issues in cloud computing<br />

Privacy is a fundamental right enshrined in the UN Universal Declaration of Human Rights. There are<br />

various forms of privacy, including ‘ the right to be left alone” and “control of information about ourselves”<br />

(Pearson, 2009). There are different types of information that need to be protected. These include any<br />

information that can be used to identify or locate an individual (e.g. name, address, credit card number<br />

and IP address). Sensitive information such as personal financial information and job performance<br />

information is considered private. Behavioural information such as viewing habits for digital content,<br />

user’s recently visited websites or product usage history need to be protected as well.<br />

Violation of privacy occurs as a result of a number of cloud dynamics. In the cloud the infrastructure is<br />

shared between organisations and is off-premise. Therefore there are threats associated with data being<br />

stored remotely and because of virtualisation. Virtualisation is a method of running multiple independent<br />

virtual systems on a less physical resource making one computer act as many, and sharing the resources<br />

of hosts across multiple environments. The cloud is also a dynamic environment. Services can be<br />

aggregated and changed dynamically by customers and service providers can change the provisioning of<br />

services anytime. Sensitive data may move around within an organisation and across organisational<br />

boundaries. Legal compliance and adequate protection has to be maintained therefore. The speed and<br />

flexibility of adjustment to vendor offerings that benefits business and provide a strong motivation for the<br />

use of cloud computing might come at the cost of compromise to the safety of data. Cloud computing<br />

enables new services to be made available in the cloud by combining other services, e.g. a ‘print on<br />

demand’ service can be provided by combining a printing service with a storage service. This procedure<br />

of service combination is typically under less control than previous service combinations carried out<br />

within traditional multi-party enterprise scenarios. There may be varied degrees of security and privacy in<br />

each of the components.<br />

Privacy risks for cloud computing may also lie in the following (Pearson, 2009):<br />

For the cloud service user: being forced or persuaded to be tracked or give personal information<br />

against their will<br />

For the organisation using the cloud: non-compliance to enterprise policies and legislation, loss of<br />

reputation and credibility<br />

For implementers of cloud platforms: exposure of sensitive information stored on the platforms<br />

(potentially for fraudulent purposes), legal liability, loss of reputation and credibility, lack of user trust<br />

and take up.<br />

For providers of applications on top of cloud platforms: legal non-compliance, loss of reputation,<br />

‘function creep’ using the personal information stored on the cloud, i.e. it might later be used for<br />

purposes other than the original cloud service intention<br />

For the data subject: exposure of personal information<br />

165


7. Regulatory issues<br />

Nomusa Dlodlo<br />

In the cloud there are a number of issues that need to be regulated. Potential physical location of data<br />

centres could be anywhere, with geography-blind distribution of applications and data. As a practical<br />

commercial matter, national regulations should be able to influence the actual deployment of cloud<br />

services in countries around the globe.<br />

Without concrete guarantees on the privacy of data held by cloud providers, the diffusion of cloud<br />

services may be hampered by the perceived risk in entrusting sensitive data to external cloud services. In<br />

the US and Europe the regulations require some cloud offerings to allow users to stipulate the country in<br />

which their data will be stored. Non-US firms whose servers are located in the US can have their<br />

information accessed by the US government under the US Patriot Act and Homeland Security Act<br />

(Carrigan, 2008). This impacts on information privacy policy.<br />

Strongly related to the notion of service level agreements and policy, is that of governance – how to<br />

manage sets of virtual resources. At the infrastructure level, applications may consist of many virtual<br />

machines, virtual storage and virtual networks. Managing these virtual missions, or virtual data centres,<br />

requires policy and enforcement from both the provider and consumer (Lee, 2010).<br />

In a private cloud, the infrastructure for implementing the cloud is controlled completely by the enterprise.<br />

Typically, private clouds are implemented in the enterprise’s data centre and managed by internal<br />

resources. A private cloud maintains all corporate data in resources under the control of the legal and<br />

contractual umbrella of the organisation. This eliminates the regulatory, legal and security concerns<br />

associated with information being processed on third party computing resources.<br />

In a public cloud however, external organisations provide the infrastructure and management required to<br />

implement the cloud. Public clouds have the disadvantage of hosting data in an offsite organisation<br />

outside the legal and regulatory umbrella of the organisation. In addition, as most public clouds leverage<br />

a worldwide network of data centres, it is difficult to document the physical location of data at any<br />

particular moment. These issues result in potential regulatory compliance issues which preclude the use<br />

of public clouds for certain organisations or business applications.<br />

According to Enki, et.al (Enki, 2010), the identified regulatory issues in the cloud are in the areas of<br />

service level agreements (SLA), service and support and performance. Cloud-computing services define<br />

an SLA as some guarantee of how much time the server, platform or application will be available. For<br />

example, a cloud provider will provide 99.99% uptime, or five minutes downtime a year, with a 10%<br />

discount on charges for any month in which that availability is not achieved. Since its infrastructure is not<br />

built to reach this uptime, it is effectively offering a 10% discount on services in exchange for the benefit<br />

of claiming that reliability. Another trick is to compute the SLA on an annualised basis. This means that<br />

customers are eligible for a service only after one year has passed. The end-user should pay close<br />

attention to the details of the SLA being provided and weigh that against what business impact it will have<br />

if the service provider misses the committed SLA and regulatory authorities should chip in to level the<br />

playing field.<br />

One of the greatest attractions of cloud computing is that it enables computing to be available to a large<br />

community. In addition, the elimination of the responsibility for physical hardware removes the need for<br />

data-centre administration staff. As a result, there is an increasing number of people responsible for<br />

production computing who do not have systems administration backgrounds, which creates demand for<br />

comprehensive cloud vendor support offerings and thereby greatly inconveniencing the consumer. Round<br />

the clock live support staff costs a great deal.<br />

The cloud is a pervasive federated network in which unregulated personal area networks and local area<br />

networks will interoperate with traditionally regulated electronic communication services. Regulators need<br />

to carefully monitor the challenges posed by these networks, taking action as necessary to regulate for<br />

technical interoperability, consumer protection, support for competition and the appearance of<br />

opportunities for the exploitation of market power.<br />

166


8. Discussion and way forward<br />

Nomusa Dlodlo<br />

This sections draws recommendations from literature on the way forward in cloud computing. These<br />

recommendations centre around the issues of policy interventions, standards, privacy and data<br />

protection, traffic and congestion management, business continuity planning, security and regulation.<br />

Any large scale deployment needs to adhere to certain standards. The cloud spans multiple industries<br />

and differs widely in application scenarios and user requirements. The standardisation of the cloud<br />

should cover common communication protocols, at, for example, the carrier level; terminal description<br />

and service discovery mechanisms and application data switching mechanisms such as technologies<br />

based on XML, SOAP and web services. The latter covers terminal, communication protocols and<br />

application specifications.<br />

Concerns over privacy and data protection are widespread. Protecting privacy must not be limited to<br />

technical solutions but encompass regulatory, market-based and socio-ethical considerations. There<br />

should be a concerted effort involving government, civil society and private sector players to protect these<br />

values. One of the hazards of shared infrastructure is that one customer’s usage pattern may affect other<br />

customers’ performance. The cloud must incorporate traffic and congestion management. This will sense<br />

and manage information flows, detect overflow conditions and implement resource reservation for timecritical<br />

and life-critical data flows.<br />

Policy interventions towards the adoption of cloud computing should include (Etro, 2010):<br />

International agreements in favour of unrestricted flow of data across borders (since data centres are<br />

located in different countries with different privacy laws, data portability remains a key issue for the<br />

diffusion of cloud computing)<br />

Agreements between government and industry leaders on a minimum set of technological standards<br />

and process standards to be respected in the provision of cloud computing services to guarantee<br />

data security and privacy and promote a healthy diffusion of new technology<br />

Expansion of broadband capacity<br />

Introduction of fiscal incentives for the adoption of cloud computing and a specific promotion in<br />

particular dynamic sectors ( for instance, governments could finance, up to a limit, the variable costs<br />

of computing for all domestic and foreign firms that decide to adopt a cloud computing solution)<br />

Some examples of cloud computing risks for the enterprise that need to be managed include (Cloud<br />

computing: business benefits, 2009):<br />

Reputation, history and sustainability of the provider to ensure reliability of service provision<br />

The cloud provider should take responsibility for information handling and be held liable for loss of<br />

confidentiality and privacy<br />

Business continuity and disaster-recovery plans must be well documented and tested<br />

Compliance to regulations and laws in different geographic regions can be a challenge for<br />

enterprises. It is critical to obtain proper legal advice to ensure that the contract specifies the areas<br />

where the cloud provider is responsible and liable for ramifications arising from potential issues.<br />

Business continuity planning is identifying core operational systems and work processes that an<br />

organisation requires in order to deliver services and products to their customers. This involves<br />

identifying key suppliers, business partners and staff. From an IT perspective this means three things<br />

(Cloud computing, 2009):<br />

Architecting your IT infrastructure and application systems to be distributed with no single point of<br />

failure<br />

Ensuring the systems have built-in data and application redundancy<br />

Having the ability for these systems to be accessed securely from any location at any time.<br />

The problem that has made business continuity and disaster recovery extremely expensive has always<br />

been the need for redundant hardware, both on-site and in remote sites. The advent of cloud computing<br />

has made the provision of dynamically scalable and virtualised resources widely and cheaply available.<br />

Security is one of the largest concerns for the adoption of cloud computing. Seven risks a cloud user<br />

should raise with vendors before committing are (Foster, 2010):<br />

167


Nomusa Dlodlo<br />

1. Privileged user access: sensitive data processed outside the enterprise needs the assurance that<br />

they only accessible and propagated to privileged users<br />

2. Regulatory compliance: A customer needs to verify if a cloud provider has external audits and<br />

security certifications and if their infrastructure complies with some regulatory security requirements<br />

3. Data location: since a customer will not know where her data will be stored, it is important that the<br />

cloud provider commit to storing and processing and processing data in specific jurisdictions and to<br />

obey local privacy requirements on behalf of the customer.<br />

4. Data segregation: one needs to ensure that one customer’s data is fully segregated from another<br />

customer’s data<br />

5. Recovery: it is important that the cloud provider has an efficient replication and recovery<br />

mechanism to restore data if a disaster occurs<br />

6. Investigative support: Cloud services are especially difficult to investigate, if this is important for a<br />

customer, then such support needs to be ensured with a contractual commitment<br />

7. Long-term viability: your data should be viable even if the cloud provider is acquired by another<br />

company.<br />

To avoid customer lock-in, customers would want to see open/standard APIs. Cloud users must not<br />

entrust mission-critical applications to cloud service providers so as to avoid outages when cloud service<br />

providers go out of business. They should also keep backups of the applications and data on premises.<br />

They should also secure favourable service-level agreements from the cloud service provider (Kim, 2009)<br />

9. Conclusion<br />

Cloud computing has transformed the knowledge society by offering enterprises the opportunities to<br />

decouple their IT needs and their infrastructure. It has therefore given rise to new business models and<br />

given the opportunity for those enterprises that would not have had the resources to compete in the<br />

knowledge society a renewed opportunity. As a result the area of cloud computing deserves to be given<br />

attention and further developed. This research does exactly that by contributing to the advancement of<br />

knowledge in that field.<br />

References<br />

Binning, D., Top five cloud computing security issues, [online],<br />

http://www.computerweekly.com/Articles/2010/01/12/235782/Top-five-cloud-computing-security-issues.htm<br />

Carrigan, M., Alex, T., Ward, C., The US Patriot Act deconstruction, Civil Liberties and Patriotism, Journal of<br />

Business and Economic Research, Vol. 6, No. 3., pp. 19-30, 2008<br />

Chow, R., Golle, P., Jakobsson, M., Masuoka, R., Molina, J., Controlling data in the cloud: outsourcing computation<br />

without outsourcing control, CCSW’09, Novembr 13, 2009, Chicago, Illinois, USA, pp. 85-90.<br />

Cloud computing as a business continuity plan, [online], http://www.tectonic.co.za/2009/05/cloud-computing-as-abusiness-continuity-plan/<br />

Enki, D., Why cloud computing will never be free, ACMQUEUE, Distributed Computing, pp. 1-10, 2010.<br />

Etro, F., The economics of cloud computing, [online], http://www.voxeu.org/index.php?q=node/4671<br />

Foster, I., Zhao, Y., Raicu, I., Lu, S., Cloud computing and grid computing 360-degree compared<br />

Gurav, U., Shaikh, R., Virtualisation – a key feature of cloud computing, International <strong>Conference</strong> and Workshop on<br />

Emerging Trends in Technology (ICWET 2010), pp. 227-229, TCET, Mumbai, India, 2010<br />

Hidden risks of cloud computing, [online], http://lifehacker.com/5325169/the-hidden-risks-of-cloud-computing<br />

How cloud computing can transform business, [online],<br />

http://blogs.hbr.org/cs/2010/06/business_agility_how_cloud_com.html<br />

Kim, W., Kim, S. D., Lee, e., Lee, S., Adoption issues for cloud computing, proceedings of MoMM2009, December<br />

14-16, 2009, Kuala Lumpur, Malaysia, pp. 2-5.<br />

Kushida K.E., Breznitz,D., Zysman, J., Cutting through the fog: understanding the competitive dynamics in cloud<br />

computing, The Berkeley Roundtable on the International Economy (BRIE) Working Paper 190 (Beta), May 1,<br />

2010<br />

Lee, C.A., A perspective on scientific cloud computing, HPDC 2010, June 20-25, 2010, Chicago, USA.<br />

Legal implications of cloud computing – Part One (the Basics and Framing the Issues), [online],<br />

http://www.llrx.com/features/cloudcomputing.htm<br />

Legal issues associated with cloud computing, [online], http://www.labnol.org/internet/cloud-computing-legalissues/14120/<br />

Pearson, S., Taking account of privacy when designing cloud computing services, CLOUD’09, ICSE’09 Workshop,<br />

pp. 44-52, Vancouver, Canada, May 23, 2009.<br />

The wisdom of clouds, [online], http://blog.jamesurquhart.com/2008/08/cloud-computing-bill-of-rights.html<br />

Zhang, Q., Cheng, L., Boutaba, R., Cloud computing: state-of-the-art and research challenges, Journal of Internet<br />

Serv Appl, Vol 1, pp. 7-18, 2010<br />

168


Interoperability Monitoring for eGovernment Service Delivery<br />

Based on Enterprise Architecture<br />

Badr Elmir 1 , Nabil Alrajeh 2 and Bouchaib Bounabat 1<br />

1 Université Mohammed V – Souissi, Morocco<br />

2 King Saud University, Saudi Arabia<br />

b.elmir@daag.finances.gov.ma<br />

nabil@ksu.edu.sa<br />

bounabat@ensias.ma<br />

Abstract: Public administration has to prepare itself to deliver fully integrated eGovernment services. This delivery<br />

often requires cooperation via business processes interoperability across two or more departments. In this context,<br />

public departments and agencies need to implement interoperability using enterprise architecture techniques to<br />

structure business processes, and service oriented models to achieve their integration. Thus, it’s quite interesting to<br />

adopt enterprise architecture paradigm and techniques to analyse, track and control the evolution degree of<br />

processes interoperability from the existing “as-is” state to the future “to-be” state. The present paper proposes a<br />

periodic monitoring approach based on an assessment method which considers three main aspects of interoperation:<br />

1. Potentiality, reflecting the preparation to interoperate. The objective is to foster interoperation readiness by<br />

eliminating barriers that may obstruct the interaction. 2. Compatibility, referring to interoperation implementation<br />

through adequate engineering process. It aims to study the relation between the external interfaces of processes and<br />

the surrounding environment in order to ensure effective interaction. 3. Performance efficiency, focusing on<br />

monitoring operational performance. It consists of the availability assessment of the communication infrastructure<br />

and the supporting system in general. It considers also end users satisfaction of interoperation in use. The proposed<br />

method supporting tool, (IMT) for interoperability monitoring tool, assesses interoperability degree periodically<br />

through five steps: (i) Delineating the scope of interoperation; (ii) Quantifying the interoperation potentiality; (iii)<br />

Calculating the compatibility degree; (iv) Evaluating the operating performance; (v) Aggregating the degree of<br />

interoperability. In addition to its capacity to track the evolution of interoperation degree in time, the IMT measures<br />

the required effort to reach a planned degree of interoperability. Finally, to better illustrate how to use the proposed<br />

interoperability monitoring approach, we present a practical example of integrated public eService. It’s a citizen<br />

oriented eService proposed by a public hospital that offers special fees for persons covered by social security<br />

insurances. It includes government to business collaboration and government to government one.<br />

Keywords: integrated public eService, enterprise architecture, interoperability assessment, periodic monitoring, and<br />

eHealth<br />

1. Introduction<br />

Public administration has to prepare itself to provide fully integrated online services for citizens and<br />

businesses. In this context, horizontal cooperation in the public domain is a key enabler for eGovernment.<br />

Indeed, the delivery of most useful online governmental services often requires cooperation between two<br />

or more public administrations or agencies. This cooperation starts from simple information exchange<br />

and can reach business processes interoperability among public departments (Klischewski 2004).<br />

The present work focuses on monitoring interoperability between automated business-processes<br />

involved in the provision of an integrated public eService. The studied processes may be located within a<br />

single organization or across a group of public partners. Therefore, the proposed approach is based on a<br />

five step measurement method (Elmir 2010b), and takes into account three main aspects:<br />

Interoperability maturity level of the environment surrounding the studied eService.<br />

Compatibility degree between the external interfaces of the involved business processes.<br />

Operational performance of the support systems used to provide the online service.<br />

The objective of this work is to: (1) Identify the most important characteristics of interoperability used to<br />

deliver integrated public services. This work proposes a set of criteria used to assess interoperability in<br />

this context considering all aspects of collaboration. (2) Describe a monitoring approach of<br />

interoperability. This allows to know what is needed to reach a desired level of interoperability.<br />

In this article, the second section is devoted to eGovernment system interoperability. The third section<br />

presents the assessment method based on a set of IT indicators for interoperability measurement. The<br />

fourth section proposes the monitoring approach model adopted in this study. This section presents also<br />

the platform developed to support interoperability monitoring in the context of integrated public eServices<br />

169


Badr Elmir et al.<br />

delivery. Finally, a sample case study is presented. It is about a citizen oriented eHealth service that<br />

encloses government-to-government and government-to-business collaboration.<br />

2. Interoperability in eGovernment system<br />

2.1 eGovernment service delivery<br />

Government-to-government collaboration is a key factor in the success of almost all eGovernment<br />

strategies. In this context, Public online eServices target multiple groups of potential customers that could<br />

be organized into four main categories (Vintar 2002):<br />

Government-to-government services (Chen 2009): generally known by e-administration, it refers to<br />

public interdepartmental coordination and collaboration in order to deliver electronic services.<br />

Integrating the intergovernmental systems is the first step in implementing electronic government<br />

solutions.<br />

Government-to-citizens services (Peristeras 2009): this field implements the relationship between the<br />

government and the citizen, and provides communication information and essential services that<br />

interest individuals or residents in order to ease their life. It represents the main aim of eServices<br />

projects.<br />

Government-to-business services (Shambour 2010): it refers to online interaction between<br />

government and the business sector. This category includes optimizing administrative procedures,<br />

public tenders, various permits, authorizations, electronic transaction services including<br />

<br />

procurements and bids; and payment services of various taxes and public charges.<br />

Government-to-employees services (Bercea 2010): it includes a set of employees and managers self<br />

services giving the ability to view and update personal, administrative, payroll and benefits<br />

information via human resources portals. These include also communication, knowledge<br />

management, and collaboration tools made available to employees to well perform their daily<br />

operations.<br />

In term of service nature, public online services have different types (Chen 2006):<br />

Informational use: in which public departments and agencies publish information to educate,<br />

entertain, influence, or reach their potential customers;<br />

Transactional use: where they support a coordinated sequence of user and system activities to<br />

provide a specific service;<br />

Operational use: when an agency provides a new mechanism for conducting business operations by<br />

integrating information systems, human intellect, and other resources into synergistic networks.<br />

In this context public administration can be viewed as a business collaboration network (BCN) (Elmir<br />

2010b). In fact, BCN enables organizations to communicate and collaborate with their customers,<br />

partners and suppliers in a productive way (Sterling 2010). This cooperation takes different forms that<br />

start from simple information exchange, and can reach business processes interoperability among<br />

independent administrations (Sun 2007) (Shishkov2009).<br />

In this sense, this paper studies more precisely the back office integration of public administration<br />

(government-to-government) in order to provide operational integrated eServices essentially those<br />

oriented to citizens and businesses. In order to preserve the autonomy of the actors, this integration<br />

takes usually the form of business process interoperability which represents an obvious prerequisite of<br />

integrated public eService delivery.<br />

2.2 Interoperability<br />

Interoperability characterises the ability, for any number of processing information systems, to interact<br />

and exchange information and services between them.<br />

To implement interoperability, public administration faces technical and semantic difficulties (Gupta 2007)<br />

but also organizational challenges (Goldkuhl 2008). Moreover, monitoring this quality is not easy on such<br />

a macroscopic level.<br />

In fact, interoperability is an information system quality that can be viewed from various perspectives.<br />

Several taxonomies have been proposed in this direction. In this sense, there are:<br />

170


Badr Elmir et al.<br />

Many levels of interoperability scope: business, process, service and data level (Guijaro 2007).<br />

Various approaches to implement this quality: integrated, federated, and unified approach (Missikoff<br />

2004).<br />

Multiple barriers could handicap interoperation: conceptual, organizational and technical barriers<br />

(Arms 2002).<br />

Different scopes of application: within the same organization, cross independent organizations<br />

(Guédria 2008),<br />

Different transactional aspects of cooperation: synchronous or asynchronous collaboration<br />

(Michelson 2006).<br />

Diverse measurement perspectives: potentiality, compatibility, performance efficiency (Chen 2008).<br />

Thus, in terms of scope, there are various levels where interoperability takes place within public<br />

administration (Guijaro 2007):<br />

Business level that refers to how to work within a business network in harmonized way in order to<br />

collaborate.<br />

Process level aims making various processes working together. In the case of a networked<br />

administration, internal processes of two departments are connected to create a common macro<br />

process.<br />

Service level is concerned with identifying, composing, and making function together with various<br />

applications.<br />

Data level refers to making synergy between different data models and heterogeneous conceptual<br />

schemas.<br />

This organization coincides with the four levels of enterprise architecture metamodel.<br />

Also, in terms of barriers, the interoperability implementation faces (Arms 2002):<br />

Conceptual barriers which are related to the syntactic and semantic problems of information to be<br />

exchanged.<br />

Organizational barriers which refer to the definition of responsibilities and authority so that<br />

interoperability can take place.<br />

Technical barriers which deal with the use of adequate protocols, languages and infrastructure in<br />

communication.<br />

In this case, interoperation compatibility check has to consider these barriers on each one of the four<br />

enterprise architecture layers cited before.<br />

2.3 Enterprise Architecture in government structure<br />

Public services users expect to perceive public administrations as a homogeneous and coherent unit in<br />

order to have a unified access to services they need. So, Public administration must be prepared to<br />

interact effectively with all the surrounding actors. This requires essentially openness and willingness to<br />

break functional, organizational and technological barriers.<br />

In this context, several governments have established Enterprise Architecture programmes (Liimatainen<br />

2007). These programmes are well known by Government Enterprise Architecture (GEA). They aim to<br />

eliminate overlapping projects, to support reuse, and to enhance interoperability between public<br />

departments. On another hand, some national strategies were limited to the single issue of interoperation<br />

and developed interoperability frameworks. They are mainly addressing technical problems by<br />

referencing the main specifications recommended to facilitate and promote cooperation between different<br />

government agencies (Guijaro 2007).<br />

At the same time, the concept of enterprise architecture (EA) attracted a lot of interest during the past<br />

decade. It aims to provide a structure for business processes and systems that supports them. It<br />

represents an information system using models in order to illustrate the interrelationship between its<br />

components and its relationship with the ecosystem.<br />

171


Badr Elmir et al.<br />

The Enterprise Architecture (EA) proposes to take an inventory of information system components by<br />

considering: (1) organization procedures, etc. (2) business process (3) IT applications, (4) technical<br />

infrastructure.<br />

Also, with the emergence of service delivery environments, independent businesses become able to<br />

collaborate in order to have beneficial results for all (Aisopos 2009). Among the main forms of<br />

cooperation, occurs integrated service providing to clients.<br />

The service-oriented interaction model implements less coupled connections between various distributed<br />

software components. The approach seeks to provide abstraction by encapsulating functionality and<br />

allowing reuse of existing services.<br />

To facilitate interoperation within a public collaboration network, usually public administration tends to<br />

adopt the enterprise architecture as a strategic choice of organization using "the service oriented"<br />

paradigm and techniques to implement and deploy services.<br />

3. Interoperability assessment method<br />

3.1 Interoperability measurement aspects<br />

Concerning measurement, Chen (2008) differentiates between the following complementary<br />

characteristics (see Figure 1):<br />

Interoperation potentiality: it is an «internal quality» of a system that reflects its preparation to<br />

interoperate. This involves identifying a set of characteristics that have an impact on communication<br />

with peer’s systems without necessarily having concrete information on them. The objective is to<br />

foster interoperability readiness by eliminating barriers that may obstruct the interaction.<br />

Interoperation compatibility: it represents an «external quality». In fact, the interaction between two<br />

support systems is ensured through an adequate engineering process.<br />

Interoperation performance: the third aspect characterizes the «quality in use». It focuses on<br />

monitoring operational performance. It consists of an assessment of the communication infrastructure<br />

availability, and the supporting system in general.<br />

Figure 1: Operational aspects of interoperability (Elmir 2010b)<br />

3.2 Interoperability indicators<br />

Many interoperability maturity models (IMM) were introduced to describe the interoperation potentiality.<br />

They are mostly inspired by the CMM/CMMI model (Chrissis 2003). Pardo (2009) Lists among others:<br />

ITIM (It Investment Management),<br />

LISI (Level of Information System Interoperability),<br />

OIMM (Organizational Interoperability Maturity Model),<br />

EIMM (Enterprise Interoperability Maturity Model),<br />

172


Badr Elmir et al.<br />

GIMM (Government Interoperability Maturity Matrix),<br />

SPICE (Software Process Improvement and Capability dEtermination).<br />

Each model adopts a specific vocabulary to express the levels of maturity. However, the models have in<br />

general five scales ranging from low to high:<br />

An organization with a low level of interoperability is characterized as working independently or in<br />

isolation from other organizations and in an ad hoc or inconsistent manner.<br />

An organization with a high level of interoperability is characterized as being able to work with other<br />

organizations in a unified or enterprise way to maximize the benefits of collaboration across<br />

organizations.<br />

In terms of compatibility and in order to dematerialise a business process and to interconnect it with its<br />

ecosystem, there is a need to study the external interfaces of its support systems. In this case, the<br />

degree of compatibility «DC» is calculated on the basis of a mapping between the underlying<br />

components and the adjacent processes.<br />

Several studies have focused on the characterization of the interoperation compatibility. Kasunic (2004)<br />

identifies for instance several indicators to describe this compatibility.<br />

The operational performance «PO» measurement is done on the basis of IT dashboards of involved<br />

public departments. It takes into account indicators as the availability score of the application servers, the<br />

quality of service of communication, and the end users degree of satisfaction about the interoperation in<br />

use. This information is collected based on surveying end users.<br />

3.3 Interoperability assessment method<br />

The present work proposes a five step method to assess interoperability needed to deliver a specific<br />

integrated public eService. These steps are as follows (see Figure 2):<br />

1. Delineating the scope of the study.<br />

2. Quantifying the interoperation potentiality.<br />

3. Calculating the compatibility degree.<br />

4. Evaluating the operating performance.<br />

5. Aggregating the degree of interoperability.<br />

Figure 2: Five steps of interoperability measurement (Elmir 2010a)<br />

173


3.4 Delineating the scope of the study<br />

Badr Elmir et al.<br />

Assessing interoperability, whether used or required, to deliver a specific integrated public eService<br />

requires the knowledge of its ecosystem.<br />

In practical terms, the study focuses on a macro business process consisting of a set of sub automated<br />

processes in independent departments. These sub processes are linked together via several interfaces<br />

identified in advance. In this case, the preliminary phase consists of identifying the context of the studied<br />

online service delivery and then lists its underlying automated processes.<br />

This step includes identifying:<br />

Public departments and agencies involved in the cooperation.<br />

Sub process within each department in order to study compatibility.<br />

Information systems that support automated business processes within each department.<br />

3.5 Quantifying the interoperation potentiality<br />

The calculation of the potential for interoperability within the k th department «PIk» requires the adoption of<br />

one of these maturity models mentioned above. The organization is classified then on one of these five<br />

levels noted IMML (for interoperation maturity model level). To identify the potential degree of<br />

interoperability, we propose then the following mapping (See Table 1):<br />

Table 1: Quantification of the maturity of the interoperability<br />

Maturity Level (IMML) Potentiality quantification<br />

1 0.2<br />

2 0.4<br />

3 0.6<br />

4 0.8<br />

5 1<br />

Within each department, the potential is calculated using the following equation:<br />

The final interoperation potentiality is given by:<br />

3.6 Calculating the degree of compatibility<br />

To assess the compatibility degree, the present work uses the compatibility matrix of Chen and Daclin<br />

(Chen 2008).<br />

The compatibility matrix (see Table 2) consists of a combination of the “interoperability levels perspective”<br />

and “interoperability barriers perspective” seen in section 2.2. In practical terms, we enumerate<br />

conceptual, technical and organisational barriers in the different layers of interoperability scope: business,<br />

process, service and data.<br />

Table 2: interoperability compatibility<br />

Conceptual Organizational Technology<br />

syntactic semantic<br />

authorities<br />

responsibilities<br />

Organization platform communication<br />

Business dc11 dc12 dc13 dc14 dc15 dc16<br />

Process dc21 dc22 dc23 dc24 dc25 dc26<br />

Service dc31 dc32 dc33 dc34 dc35 dc36<br />

Data dc41 dc42 dc43 dc44 dc45 dc46<br />

174


Badr Elmir et al.<br />

By noting the elementary degree of interoperation compatibility «dcij» (i takes values from 1..4, and j<br />

takes values from 1..6). Therefore, if the criterion in an area mark satisfaction the value 0 is assigned to<br />

dcij; otherwise if a lot of incompatibilities are met, the value 1 is assigned to dcij.<br />

The degree of compatibility «DC» is given as follows:<br />

3.7 Evaluating operating performance<br />

By Denoting:<br />

«DS» the overall availability rate of application servers.<br />

«QoS» service quality of different networks used between interacting components. QoS is<br />

represented mainly by the overall availability of networks.<br />

«TS» end users’ satisfaction level about interoperation.<br />

Given the cumulative nature of these three rates, the evaluation of operational performance is given by<br />

the geometric mean (DeFusco 2007) as the following equation:<br />

3.8 Aggregating the degree of interoperability<br />

The final calculation of the ratio characterising the interoperability process in question is by aggregating<br />

the three previous indicators using a function f defined in [0,1]3 [0,1]<br />

RatIop=f (PI, DC, PO)<br />

Given the independent nature of these three indicators, we opt for the arithmetic mean (DeFusco 2007)<br />

as follows:<br />

4. Periodic interoperability monitoring approach<br />

4.1 Interoperability monitoring approach<br />

This section shows the metamodel we propose for the interoperability monitoring approach. It includes:<br />

The exposed public integrated online service.<br />

The business processes supporting the delivery of the public online service.<br />

The connections and the compositions that exist between the involved business processes.<br />

The public departments that participate to the provision of the online service.<br />

The maturity model used in every department and their levels and the prerequisites to reach each<br />

level.<br />

The end users Satisfaction level.<br />

The enterprise architecture layers that coincide with the level of interoperation scopes.<br />

The elementary barriers that may obstruct interoperation situations.<br />

The periods within which interoperability is assessed.<br />

The metamodel serves as a basis for interoperability assessment periodically. It considers existing IT<br />

indicators within the public collaboration network like availability rate of application servers and the<br />

network. It includes end users satisfaction about used interoperation. This metamodel includes maturity<br />

score of each involved department. It references furthermore compatibility aspects on all levels of the<br />

enterprise architecture of collaboration in use.<br />

175


Badr Elmir et al.<br />

The metamodel is represented by using UML diagrams: class diagram (See Figure3).<br />

Figure 3: Interoperability monitoring Metamodel (class diagram)<br />

4.2 Periodic interoperability monitoring tool<br />

The Interoperability monitoring tool (IMT) includes three principal modules. The first one is dedicated to<br />

the interoperability assessment at a specific period. Figure 4 describes the interoperability assessment of<br />

an automated macro process within a single department which uses GIMM as maturity model. In this<br />

specific case, we notice that there are a lot of conceptual and organizational incompatibilities on a<br />

business layer.<br />

Figure 4: Screen from Interoperability monitoring tool<br />

176


Badr Elmir et al.<br />

In addition to its capacity to track the evolution of the interoperation degree periodically, the IMT gives the<br />

possibility to propose a scenario to reach a planned degree of interoperability. For instance, in the<br />

example shown on Figure 5, we plan to move the interoperability ratio calculated from 79% to 89%.<br />

IMT proposes to (i) improve the interoperability maturity to reach the fourth level, (ii) optimize the<br />

availability of involved application servers, (iii) better meet end users expectations and (iv) resolve<br />

conceptual incompatibilities.<br />

Figure 5: Planning of Interoperability optimization<br />

Interoperability monitoring tool is able to recognize prerequisites of going from a specific level of maturity<br />

to the next one.<br />

Figure 6: Sample of interoperability maturity optimization details<br />

5. Case study<br />

This section presents an illustration example on monitoring interoperability of an integrated public<br />

eService. The case in this paper consists of an online payment for health care services. It’s a citizen<br />

177


Badr Elmir et al.<br />

oriented eService proposed by a public hospital that offers special fees for persons covered by social<br />

security insurances.<br />

The studied online service is designed and implemented over an inter organizational information system<br />

within four public departments as described in figure 7. This eService includes:<br />

The identification and notification services within the public hospital system.<br />

The fee payment in national treasury department system.<br />

The administrative data validation in the national general mutual: it is oriented essentially to the civil<br />

servants.<br />

The administrative data validation in national social security fund: it is oriented to employees in the<br />

private sector.<br />

The data validation within private insurance: it is oriented to citizens who contracted a complementary<br />

insurance.<br />

Figure 7: Sample of online payment situation in eHealth system<br />

The interoperability is assessed every quarter in this collaboration network. During this period, the<br />

maturity level is brought up. The end user satisfaction is improved. The IT indicators are optimized (See<br />

Figure 8).<br />

Figure 8: Quarterly interoperability monitoring eHealth online payment<br />

178


5.1 Discussion<br />

Badr Elmir et al.<br />

The provided case study illustrates the results of interoperability monitoring within a specific collaboration<br />

context. This is provided by the interoperability monitoring tool (IMT) that collects existing IT indicators<br />

like availability score of Application servers and the end users satisfaction level. The IMT logs, in a<br />

convenient way, the interoperation incompatibilities in every architecture layer surrounding the delivery of<br />

the integrated eService.<br />

This automated method enables the establishment of an action plan that aims to improve the integrated<br />

eService delivery.<br />

The IMT is able to propose a scenario to reach a planned result for interoperability ratio. However the<br />

current version of the IMT is not yet able to propose the best scenario to achieve this result efficiently.<br />

This work is a prerequisite for several projects launched in parallel, and dealing with the applicability of<br />

control theory in the information system interoperability field. In this case, the future version of the IMT<br />

will use such techniques to enable the optimal control of interoperability.<br />

6. Conclusion<br />

The delivery of most useful public online services often requires cooperation between two or more public<br />

administrations. This cooperation takes, in general, the form of business processes interoperability<br />

among public departments. The present paper has presented a periodic monitoring approach for<br />

interoperability to deliver integrated public eServices. This approach is based on a five step assessment<br />

method that uses existing indicators within involved departments like quality maturity indicators,<br />

information technology dashboards, etc.<br />

The result of this automated method is a ratio metric enabling the measurement of this quality by taking<br />

into account three main operational aspects: interoperation potentiality, interoperation compatibility and<br />

operational performance.<br />

The proposed method supporting tool tracks the evolution of interoperation degree in time and is able to<br />

propose a suitable scenario to reach planned degrees of interoperability.<br />

References<br />

Aisopos, F., Tserpes, K., Kardara, M., Panousopoulos, G., Phillips, S. and Salamouras, S. (2009) “Information<br />

exchange in business collaboration using grid technologies, Identity in the Information Society”, Springer 2009,<br />

Volume 2, Number 2, pp. 189-204.<br />

Arms, W. Y., Hillmann, D., Lagoze, C., Krafft, D., Marisa, R., Saylor, J. et al. (2002) “Spectrum of Interoperability”:<br />

The Site for Science Prototype for the NSDL, D-Lib Magazine, 8 (1), January (2002)<br />

Bercea, L., Nemtoi, G. and Ungureanu, C. (2010), “The government of state’s power bodies by means of the<br />

Internet”, Journal of Computing, Vol. 2, Issue 2, February 2010.<br />

Chen, D. and Daclin, N. (2006) “Framework for Enterprise Interoperability”. IFAC TC5.3 workshop EI2N, Bordeaux,<br />

France.<br />

Chen, D., Vallespir, B. and Daclin, N. (2008) “An Approach for Enterprise Interoperability Measurement”. Paper read<br />

at MoDISE-EUS 2008. France.<br />

Chen, J. (2009) “Qualitative Evaluation of EGovernment Service Based on Participatory Index and Balanced<br />

Scorecard Approach”, iciii, International <strong>Conference</strong> on Information Management, Innovation Management and<br />

Industrial Engineering, Vol. 4, pp. 380-383.<br />

Chrissis, M.B., Mike, C., Sandy S. (2006) CMMI: Guidelines for Process Integration and Product Improvement.<br />

Addison-Wesley. USA.<br />

DeFusco, R., McLeavey, D.W. and Pinto J.E. (2007) Quantitative investment analysis. Vol. 2 of CFA Institute<br />

investment series, pp. 127, John Wiley and Sons, Hoboken, NJ, USA.<br />

Elmir, B. and Bounabat, B. (2010a) “Integrated Public EServices Interoperability Assessment”. International Journal<br />

of information Science and Management, Special issue, October 2010, pp 1-12.<br />

Elmir, B. and Bounabat, B. (2010b) (Elmir-b) “EService Interoperability Measurement within Business Collaboration<br />

Networks”. Paper read at MICS 2010. International <strong>Conference</strong> on Models of Information and Communication<br />

Systems. Morocco.<br />

Goldkuhl, G. (2008) “The challenges of interoperability in eGovernment: towards a conceptual refinement”. Pre-ICIS<br />

2008 SIG eGovernment Workshop, Paris.<br />

Guédria, W., Naudet, Y. and Chen, D. (2008) “Interoperability maturity models - survey and comparison-“. Lecture<br />

Notes in Computer Science, Springer Berlin / Heidelberg. Vol. 5872/2009, pp. 216-225,<br />

Guijaro, L. (2007) “Frameworks for fostering cross-agency interoperability in eGovernment”. Electronic government:<br />

information, and transformation, within the advances in management information system (AMIS) series (ME<br />

Sharp).<br />

179


Badr Elmir et al.<br />

Gupta, K.M., Zang, M., Gray, A., Aha, D.W., and Kriege, J. (2007) “Enabling the interoperability of large-scale legacy<br />

systems”. (Technical Note AIC-07-127). Washington, DC: Naval Research Laboratory, Navy Center for Applied<br />

Research in Artificial Intelligence.<br />

Kasunic, M. and Anderson W. (2004) “Measuring Systems Interoperability: Challenges and Opportunities”. Technical<br />

Note CMU/SEI-2004-TN-003. Carnegie Mellon University, Pittsburgh.<br />

Klischewski, R. (2004) “Information Integration or Process Integration? How to Achieve Interoperability in<br />

administration”. In Traunmüller, R., Electronic Government: Third International <strong>Conference</strong> EGOV 2004, Lecture<br />

Notes in Computer Science 3183, Berlin: Springer Verlag, pp.57-65.<br />

Liimatainen, K., Hoffmann, M. and Heikkilä, J. (2007) “Overview of Enterprise Architecture work in 15 countries”.<br />

Ministry of Finance, State IT Management Unit, Research reports 6b/2007.<br />

Michelson, B. (2006) “Event-Driven Architecture Overview”. Technical Report. Patricia Seybold Group. Boston, USA.<br />

Missikoff, M., Schiappelli, F. and D'Antonio F. (2004) “State of the Art and State of the Practice Including Initial<br />

Possible Research Orientations”. Deliverable D8.1. <strong>European</strong> Commission, INTEROP Network of Excellence.<br />

Pardo, A. and Burke B. (2009) “Improving government interoperability. A capability framework for government<br />

managers”. Technical Report. Center for technology in government. University at Albany, NY, USA.<br />

Peristeras, V., Mentzas, G., Tarabanis, K.A. and Abecker, A. (2009) “Transforming EGovernment and E-participation<br />

through IT”, IEEE Intelligent Systems, Vol. 24, no. 5, pp. 14-19, Sep./Oct. 2009.<br />

Shambour, Q. and Lu, J. (2010) “A Framework of Hybrid Recommendation System for Government-to-Business<br />

Personalized EServices”, Seventh International <strong>Conference</strong> on Information Technology, pp. 592-597.<br />

Shishkov, B., Sinderen, M. and Verbraeck, A. (2009) “Towards flexible inter-enterprise collaboration: a supply chain<br />

perspective”, Lecture Notes in Business Information Processing, 1, Volume 24, Enterprise Information Systems,<br />

III, pp. 513-527.<br />

Sterling (2010) Optimize and Transform Your Business Collaboration Network, Sterling Commerce white paper.<br />

Sun, H.H., Huang S. and Fan, Y. (2007) “SOA-Based Collaborative Modeling Method for Cross-Organizational<br />

Business Process Integration”, Lecture Notes in Computer Science, 2007, Volume 4537, Advances in Web and<br />

Network Technologies, and Information Management, pp 522-527.<br />

Vintar, M. and Leben, A. (2002) “The Concepts of an Active Life-event Public Portal”. In Proceedings of the First<br />

International <strong>Conference</strong> on Electronic Government. Springer-Verlag, Berlin, pp. 383-390.<br />

180


Enterprise Resource Planning Implementation Differences<br />

Within the Same Methodology - Case Study From West<br />

Europe and Turkey<br />

Turan Erman Erkan<br />

At l m University, Ankara, Turkey<br />

ermanerk@atilim.edu<br />

Abstract: Enterprise Resource Planning (ERP) systems are vital for competitive edge in today’s business world.<br />

ERP is an integrated information system that serves all departments within an enterprise. Therefore, business<br />

processes should be optimised before ERP implementation in order to catch a perfect implementation. Before year<br />

2000 all the giant firms realised their ERP projects and after that best practises took place. After those trial and error<br />

based experiences, ERP vendors need to form an ERP implementation methodology. Big ERP vendors developed<br />

implementation methodologies, they were almost the same; starting from project preparation to selection,<br />

implementation to maintenance and control. Those project methodologies improved the success rate of ERP<br />

implementations. This research is both empirical and qualitative. In fact it consists of two monographic researches.<br />

They are both about ERP implementations in the same sector. One of the firms is a national one and other one is a<br />

multinational one. Both implementations done by the same consultant team with the same methodology, so the<br />

evaluation of the project is more objective than any other questionnaire based ones, which are filled by different<br />

implementation teams. The findings of the comparison are surprisingly different from each other within the<br />

implementations, although they both used the same methodology. Basic steps of the ERP implementation such as:<br />

project preparation, business blueprint, realization, final preparation and go live & support differs from each other<br />

both in national and multinational firms. The major difference is seen in the first steps which are project preparation<br />

and business blueprint. Multinational firm seems to have more chance than the national one in achieving<br />

organisational efficiency through successful ERP implementation. The research outcome is useful for professionals<br />

running implementation projects and those making decisions on ERP implementation. The results can also be used<br />

by practitioners managing ERP projects in order to avoid from implementation methodology illusion.<br />

Keywords: enterprise resource planning, ERP project management, ERP implementation, cultural differences<br />

1. Introduction<br />

The business environment is dramatically changing. Companies today face the challenge of increasing<br />

competition, expanding markets, and rising customer expectations. This increases the pressure on<br />

companies to lower total costs in the entire supply chain, shorten throughput times, drastically reduce<br />

inventories, expand product choice, provide more reliable delivery dates and better customer service,<br />

improve quality, and efficiently coordinate global demand, supply, and production (Umble, 2003).<br />

Therefore, the changes in the business perspectives, goals objectives and strategies are pressuring on<br />

organizations and on their structure to be upgraded in spite of their cultures and values. Knowledge<br />

sharing is not a competitive advantage anymore the challenge is in knowledge management.<br />

As the business world evolved it was no longer adequate for companies to merely offer their goods for<br />

sale, in order to stay viable they had to keep their competitive advantage (Tersine 2003).<br />

in the ’60s industry concentrated on how to produce more (quantity),<br />

in the ’70s how to produce it cheaper (cost)<br />

in the ’80s how to produce it better (quality)<br />

in the ’90s how to produce it quicker (lead time)<br />

in the 21st century how to offer more (service<br />

ERP systems have been considered an important development in the corporate use of information<br />

technology in the 1990s, enhancing organizational cross-functional efficiency and effectiveness through<br />

the seamless integration of all the information flowing through a company (Davenport, 1998).<br />

ERP is the business backbone. It is a cross-functional enterprise system that integrates and automates<br />

many of the internal business processes of a company, particularly those within the manufacturing,<br />

logistics, distribution, accounting, finance, and human resource functions of the business. Thus, ERP<br />

serves as the vital backbone information system of the enterprise, helping a company achieve the<br />

efficiency, agility, and responsiveness required to succeed in a dynamic business environment<br />

(Davenport, 1998) and (Olson, 2004). ERP software typically consists of integrated modules that give a<br />

181


Turan Erman Erkan<br />

company a real-time cross-functional view of its core business processes, such as production, order<br />

processing, and sales, and its resources, such as cash, raw materials, production capacity, and people.<br />

However, properly implementing ERP systems is a difficult and costly process that has caused serious<br />

business losses for some companies, which underestimated the planning, development, and training that<br />

were necessary to reengineer their business processes to accommodate their new ERP systems. ERP<br />

literature suggest that an ERP system alone cannot improve the company performance unless an<br />

organization restructures its operational processes, and this is generally accomplished through business<br />

process reengineering (Bingi, Sharma, and Godla 2003, Olson 2004 and Davenport 1998).<br />

1.1 Enterprise resource planning project management<br />

In order to study ERP implementation projects, some researchers are using a critical success factors<br />

(CSFs) approach (Esteves and Pastor 2001, Nah et al 1999). However, little has been done in relation to<br />

the management and the operationalization of these CSFs. Project evaluation is critical to the<br />

understanding, control and monitoring of the CSFs of an ERP implementation project. ERP project<br />

success is influenced by a large number of factors, and most of the times it is difficult to measure them<br />

objectively. Adequate Business Process Redesign (BPR) is one of the most cited CSFs in ERP<br />

implementation projects (e.g. Bancroft et al. 1998, Bingi et al. 1999, Holland et al. 1999, Nah et al. 2001).<br />

Comprehensive BPR is related with the alignment between business processes and the ERP business<br />

model and associated best practices (Esteves and Pastor 2000). As Valiris and Glykas (1999) mention, in<br />

the literature there has been some confusion regarding the use of terms like reengineering, process<br />

improvement and redesign. They suggest that reengineering is synonymous to radical change and<br />

process improvement to incremental change and that both, reengineering and process improvement are<br />

included in the definition of redesign.<br />

This CSF stream focuses on the details associated with implementing ERP systems and their relative<br />

success and cost. Specifically, according to Motwani et al (2005); the articles in this stream include topics<br />

such as the implementation procedures (Abdinnour et al 2003, Al-Mudimigh et al 2001, Mandal and<br />

Gunasekaran 2003 and Umble 2003), critical success factors (Al-Mashari et al 2003, Nah et al 2001,<br />

Clemmons and Simon 2001, Davison 2002 and Hong and Kim 2002), pitfalls and complexities in ERP<br />

implementation (Ip et al 2002, McAlary 1999 and Ribbers and Schoo 2002), and successful strategies for<br />

effective ERP implementation (Aladwani 2001, Cliffe 1999, Markus et al 2000a and 2000b, Scheer and<br />

Habermann 2000, Willis and Brown 2002 and Mabert, Soni and Venkataramanan 2003).<br />

Based on the preceding review of the literature done by Motwani (2005) and also on the research by<br />

Akkermans and van Helden (2002), Grabski, Leech and Lu (2001), and Somers and Nelson (2001),<br />

Grabski, Stewart, and Leech (2007) developed a list of ERP implementation controls,<br />

Business process reengineering<br />

Consultants' involvement<br />

Top management support<br />

Active steering committee<br />

Knowledgeable project team<br />

Close working relationship between the project team and consultants<br />

Detailed requirements specification<br />

Detailed implementation plan<br />

Frequent communication with the users<br />

Managing people<br />

User involvement<br />

Training<br />

Involvement of internal audit<br />

System testing prior to implementation<br />

Close monitoring after implementation<br />

Change management and transition management<br />

182


Develop users' project ownership<br />

In-depth, up front project planning<br />

Project management skills<br />

Project sponsor from top management<br />

Clearly identified objectives<br />

Specified measures of success<br />

Ways to manage risk<br />

Turan Erman Erkan<br />

Detailed tracking of actionable items by internal audit<br />

Monthly internal audit reports on project risk items to steering committee.<br />

Therefore, both ERP vendors and clients start caring about those controls and improve their<br />

implementation methodologies.<br />

1.2 SAP ASAP methodology<br />

The ASAP solution was developed to ensure the successful, on-time delivery of a project. SAP delivers<br />

the Accelerated SAP (ASAP) methodology for project management and system implementation.<br />

Developed by SAP to optimize the success of implementing the SAP Business Suite, ASAP streamlines<br />

the implementation by providing templates, methods, tools, and accelerators that have been built on the<br />

success of thousands of previous SAP implementations (SAP, 2001)<br />

The ASAP Roadmap provides the methodology for implementing and continuously optimizing your SAP<br />

System. It divides the implementation process into five phases and offers detailed Project Plans to assist<br />

you (in Microsoft Project format). The documentation stored at each level of the Roadmap tree structure<br />

contains recommendations on implementing your SAP System and links to helpful tools and accelerators<br />

(Niefert 2009, Portougal and Sundaram 2006)<br />

The implementation of your SAP System covers the following phases:<br />

Project Preparation<br />

In this phase you plan your project and lay the foundations for successful implementation. It is at this<br />

stage that you make the strategic decisions crucial to your project:<br />

Define your project goals and objectives<br />

Clarify the scope of your implementation<br />

Define your project schedule, budget plan, and implementation sequence<br />

Establish the project organization and relevant committees and assign resources<br />

Business Blueprint<br />

In this phase you create a blueprint using the Question & Answer database (Q&Adb), which documents<br />

your enterprise’s requirements and establishes how your business processes and organizational<br />

structure are to be represented in the SAP System. You also refine the original project goals and<br />

objectives and revise the overall project schedule in this phase.<br />

Realization<br />

In this phase, you configure the requirements contained in the Business Blueprint. Baseline configuration<br />

(major scope) is followed by final configuration (remaining scope), which can consist of up to four cycles.<br />

Other key focal areas of this phase are conducting integration tests and drawing up end user<br />

documentation.<br />

Final Preparation<br />

183


Turan Erman Erkan<br />

In this phase you complete your preparations, including testing, end user training, system management,<br />

and cutover activities. You also need to resolve all open issues in this phase. At this stage you need to<br />

ensure that all the prerequisites for your system to go live have been fulfilled.<br />

Go Live & Support<br />

In this phase you move from a pre-production environment to the live system. The most important<br />

elements include setting up production support, monitoring system transactions, and optimizing overall<br />

system performance.<br />

2. Case study: ERP implementations<br />

A case study approach was employed to identify the projects steps that were held differently among the<br />

national and international firm projects. The criterion used to select the case study companies was that<br />

each of the case studies should use ERP software from the same vendor. Data was collected primarily<br />

through questionnaires and observations.<br />

This case study consists of two ERP implementations. In fact, this a monographic study of two distinct<br />

firms. Firms are from Turkey because of confidentiality the sector would not be shared. one is<br />

multinational the other one is national. The purpose of the case study is to examine the ERP<br />

implementation method difference between the national and multinational firms.<br />

Both enterprises implemented the same ERP with same modules as it can be seen from Table I. The<br />

author also participated to both of the projects as a BPR and ERP consultant. Therefore, data collected<br />

from the original source. In case of confidentiality the enterprises would take place as Firm A<br />

(international) and Firm B (national)<br />

Table 1: ERP modules used in firms<br />

ERP Module Firm A Firm B<br />

FI Financial Accounting X X<br />

CO Controlling X X<br />

TR Treasury X X<br />

IM Investment Management X X<br />

PP Production Planning X X<br />

MM Materials Management X X<br />

SD Sales and Distribution X X<br />

QM Quality Management X X<br />

WM Warehouse Management X X<br />

PM Plant Maintenance X X<br />

CS Customer Service X X<br />

PS Project System X X<br />

HR Human Resources X X<br />

In Firm A implementation lasted 14 months and in Firm B implementation lasted 8 months. In both firms<br />

ASAP methodology was used. As seen from Table 1 same modules of ERP had been implemented by<br />

the same project team. There is something like 50% difference in implementation time of projects. Thus 6<br />

months difference is mainly because of project preparation and business blueprint steps. Those initial<br />

and vitals steps considered seriously in the international firm rather than the national one.<br />

Table 2 indicates the grades of the project evaluation that was realized by consultant team out of 10 for<br />

each project step. Average and standard deviation of those grades calculated for each step of ASAP<br />

methodology. According to Table 2, in all steps of ASAP methodology there is an obvious performance<br />

evaluation difference between multinational and national firms.<br />

The biggest difference is in project preparation (9.90 to 5.80) and business blueprint steps (9.80 to 6.00)<br />

which are the essential steps of an ERP project. Starting from realization (9.00 to 7.10) and continuing<br />

with final preparation (9.80 to 6.80) and go live & support (9.90 to 8.10) the difference between national<br />

and international implementation reduce. Overall average of five steps (project preparation, business<br />

184


Turan Erman Erkan<br />

blueprint, realization, final preparation, go live & support) are 9,68 and 6.76 respectively in international<br />

and national firms. In other words simply one can easily say that implementation quality is 96.8% in<br />

international firm and 67.6% in national firm.<br />

Table 2: ERP project steps evaluation realized by consultant team<br />

Firm A Firm B<br />

ASAP Steps Average St. D.* Average St. D.*<br />

Project Preparation 9.90 0.32 5.80 1.03<br />

Business Blueprint 9.80 0.42 6.00 0.67<br />

Realization 9.00 0.82 7.10 0.88<br />

Final Preparation 9.80 0.42 6.80 0.79<br />

Go Live & Support 9.90 0.32 8.10 0.74<br />

Overall Average 9.68 6.76<br />

St.D.*: Standard Deviation<br />

3. Conclusion<br />

This research attempted to answer the question. ‘‘Is there a difference in implementing ERP via same<br />

project methodology between national and international firms?” The answer was quite surprising; primary<br />

steps of the ERP implementation such as: project preparation, business blueprint, realization, final<br />

preparation and go live & support differ from each other both in national and multinational firms. The<br />

major difference is seen in the first steps which are project preparation and business blueprint. Therefore,<br />

multinational firm seems to have more probability than the national one in achieving a successful ERP<br />

implementation. Since the questions were asked to the ERP consultants that they participated to the<br />

project, there was no problem about the perception of the questions and therefore answers were<br />

objective.<br />

Although this research is a monographic one, it indicates the project management difference between<br />

national and international firms in Turkey. As discussed in the previous sections both project period (14 to<br />

8 months) and evaluated performance (9.68 to 6.76) differ among international and national firms.<br />

National firms usually do not have budget and time for business process reengineering, organizational<br />

change management and so on.<br />

In conclusion, there is a recordable ERP project implementation difference between multinational and<br />

national firms although they use the same methodology. The difference is mainly on the way of doing<br />

business, the planning method and application. This might be because of cultural differences and<br />

business circumstances. National firms’ ERP projects are mostly about configuration of the software to<br />

the enterprise not a tailor made customization for them to achieve organisational efficiency through<br />

successful ERP implementation. National firms need to redesign their organizations, reengineer their<br />

processes, enforce employees in order to increase their chance of ERP well implementation.<br />

The research outcome is useful for professionals running implementation projects and those making<br />

decisions on ERP implementation. The results can also be used by practitioners managing ERP projects<br />

in order to avoid from implementation methodology illusion. This study was subject to further limitations<br />

due to the measurements of cultural effects and economically readiness of firms that could be a subject<br />

of an other study.<br />

References<br />

Abdinnour-Helm, S. Lengnick-Hall, M. and Lengnick-Hall, C. (2003) “Pre-implementation attitudes and Organizational<br />

Readiness for Implementing an Enterprise Resource Planning System”, <strong>European</strong> Journal of Operational<br />

Research 146 (2) pp 258-273.<br />

Akkermans, H. and K. Helden, K. (2002) “Vicious and virtuous cycles in ERP implementation: a case study of<br />

interrelations between success factors” <strong>European</strong> Journal Information Systems, 11 (1) pp 35–46.<br />

Aladwani, A. (2001) “Change management strategies for successful ERP implementation”, Business Process<br />

Management 7 (3) pp 266-275<br />

Al-Mashari, M. Al-Mudimigh, A. and Zairi, M. (2003) “Enterprise Resource Planning: A Taxonomy of Critical Factors”,<br />

<strong>European</strong> Journal of Operational Research 146 (2) pp 352-364.<br />

Al-Mudimigh, A. Zairi, M. and Al-Mashari, M. (2001) “ERP Software Implementation: an Integrative Framework”,<br />

<strong>European</strong> Journal of Information Systems 10 (4) pp 216-226.<br />

Bancroft, N. Seip H.and Sprengel A. (1998) "Implementing SAP R/3", 2nd ed., Manning Publications.<br />

185


Turan Erman Erkan<br />

Bingi, P. S. Sharma M. K., and Godla J. (1999) “Critical issues affecting an ERP implementation”. Information<br />

Systems Management,16(3) pp7–14.<br />

Clemmons, S. and Simon, S. (2001) “Control and Coordination in Global ERP Configuration”, Business Process<br />

Management Journal 7 (3) pp 205-215.<br />

Cliffe, S. (1999) “ERP implementation”, Harvard Business Review 77 (1) pp 16-17.<br />

Davenport, T.H. (1998) “Putting the Enterprise into the Enterprise System” Harvard Business Review 76 (4), pp.<br />

121–131.<br />

Davison, R. (2002) “Cultural complications of ERP”, Communications of the ACM 45 (7) pp109-111<br />

Esteves, J. and Pastor, J. (2000) "Towards the Unification of Critical Success Factors for ERP implementations",<br />

10th Annual BIT conference, Manchester, UK.<br />

Esteves, J. and Pastor, J. (2001) "Analysis of Critical Success Factors Relevance along SAP Implementation Phases<br />

", Americas <strong>Conference</strong> on Information Systems.<br />

Esteves, J. Pastor, J. and Casanovas, J. (2002) "Monitoring Business Process Redesign in ERP Implementation<br />

Projects", Americas <strong>Conference</strong> on Information Systems (AMCIS), Dallas (USA).<br />

Grabski, S. V. Leech, A. and Lu, B. (2001) “Risks and Controls in the Implementation of ERP Systems”. Int J Digit<br />

Account Res;1(1) pp 51–78.<br />

Grabski, S. V. Stewart, A. and Leech,A. (2007) “Complementary Controls and ERP Implementation Success”<br />

International Journal of Accounting Information Systems 8 pp 17–39.<br />

Holland, C. P. Light, B.and Gibson, N. (1999) "A Critical Success Factors Model for Enterprise Resource Planning<br />

Implementation", <strong>European</strong> <strong>Conference</strong> on Information Systems, pp. 273-279.<br />

Hong, K. and Kim, Y. (2002) “The critical success factors for ERP implementation: an organizational fit perspective”,<br />

Information and Management (1) pp 25-40<br />

Ip, W. Chau, K. and Chan, S. (2002) “Implementing ERP through continuous Improvement”, International Journal of<br />

Manufacturing 4 (6) pp 465-478<br />

Mabert, V. Soni, A and Venkataramanan, M. (2003) “Enterprise Resource Planning: Managing the Implementation<br />

Process”, <strong>European</strong> Journal of Operational Research 146 (2) pp 302-314<br />

Mandal, P. and Gunasekaran, A. (2003) “Issues in implementing ERP: a case study”, <strong>European</strong> Journal of<br />

Operational Research 146 (2) pp 274-283<br />

Markus, M. Axline, S. Petrie, D. and Tanis, C. (2000b) “Learning from adopters’ experiences with ERP: problems<br />

encountered and success achieved”, Journal of Information Technology 15 (4) pp 245–266.<br />

Markus, M. Tanis, C. and Fenema, P. (2000a) “Multisite ERP implementations”, Communications of the ACM 43 (4)<br />

pp 42–46.<br />

McAlary, S. (1999) “Three pitfalls in ERP implementation”, Strategy and Leadership 27 (6) pp 49-53.<br />

Motwani, J, Subramanian, R and Gopalakrishna, P (2005) “Critical factors for successful ERP implementation:<br />

Exploratory findings from four case studies”, Computers in Industry Volume 56, Issue 6, pp 529-544<br />

Nah, F. Lau, J. and Kuang, J. (2001) "Critical Factors for Successful Implementation of Enterprise Systems",<br />

Business Process Management Journal, 7(3), pp. 285 -296.<br />

Nah, F. Lee-Shang J. Lau, J. and Kuang, J. (2001) “Critical Factors for Successful Implementation of Enterprise<br />

Systems”, Business Process Management Journal 7 (3) pp 285-296.<br />

Niefert, W. (2009) SAP Business ONE Implementation, Packt Publishing, Birmingham.<br />

Olson, L. (2004) Managerial Issues of Enterprise Resource Planning Systems, McGraw Hill, NY,.<br />

Portougal, V. and D. Sundaram, (2006) Business processes: Operational Solutions for SAP Implementation, IRM<br />

Press London<br />

Ribbers, P. and Schoo, K. (2002) “Program management and complexity of ERP implementations”, Engineering<br />

Management Journal 14 (2) pp 45-54.<br />

SAP (2001), ASAP Methodology service.sap.com/asap, [Accessed 7.12.2001]<br />

Scheer, A. and Habermann, F. (2000) “Making ERP a Success”, Communications of the ACM 43 (4) pp 57–61.<br />

Somers, T. and Nelson, K. G. (2001) “The impact of critical success factors across stages of enterprise resource<br />

planning implementations” Proceedings of the 34th Hawaii International <strong>Conference</strong> on Systems Science.<br />

Tersine R., (2003) Open University, Available: www.ou.edu/class/tersine.<br />

Umble, E. J. Haft, R. R. Umble, M. M. (2003) “Enterprise resource planning: Implementation procedures and critical<br />

success factors” <strong>European</strong> Journal of Operational Research 146 pp. 241–257.<br />

Valiris, G. and Glykas, M. (1999) “Critical Review of Existing BPR Methodologies”, Business Process Management<br />

Journal, 5(1), pp. 65-86.<br />

Willis, T. and Brown, A. (2002) “Extending the value of ERP”, Industrial Management and Data Systems 102 (1/2) pp<br />

35-38<br />

186


A Meta-Analysis of how the Adoption of Innovative Web 2.0<br />

Tools Like Digital Business Ecosystems can Lead to Improved<br />

SME Collaboration<br />

Francesc Estanyol<br />

University of Edinburgh Business School, Edinburgh, UK<br />

francesc.casals@ed.ac.uk<br />

Abstract: This work studies the increasing role of Web 2.0 tools as enabler of SME collaboration. The author analyses<br />

how innovative Information and Communication Technologies (ICT) like Digital Business Ecosystems (DBE)<br />

could help small and medium enterprises (SMEs) to reduce the gap with large corporations. Nowadays Web 2.0 is<br />

permitting to perform some tasks unimaginable only few years ago. Individuals have evolved from a passive to an<br />

active role and currently are the main actors creating Internet content. However, compare to individuals, business still<br />

in the first stage of Internet revolution and are not taking profit of all the possibilities of Web 2.0. Most of them are<br />

using these tools only to improve existing procedures but not for creating new ways of doing business or interact with<br />

other SMEs. In this scenario, DBE has emerged as a new paradigm to solve some of these problems and to permit<br />

businesses behave in a similar way to what individual users are doing nowadays. In this work, after an analysis of the<br />

nature of co-operative behaviour and the importance of Web 2.0 facilitating its emergence by promoting selforganisation<br />

of their participants, the author evaluates this new paradigm in order to determine if it is a valid approach<br />

for SMEs. Finally, based on previous works of the author about the needs of SMEs, the study concludes with<br />

the definition of MADBE, an innovative multi-agent Digital Business Ecosystems inspired by natural ecosystems<br />

which aims to facilitate collaboration between SMEs.<br />

Keywords: SME, collaboration, multi-agents, digital, ecosystems, business<br />

1. Introduction<br />

The field of SME collaboration has become very popular from the 90s. Terms such as networking, cooperation<br />

or alliances have been widely used in literature in works like (Hoffmann and Schlosser 2001),<br />

(Varamaki and Vesalainen 2003) or (Narula 2004) but in spite of some studies like (Coombs et al. 1996)<br />

have demonstrated the advantages of adopting collaboration and co-operation approaches for improving<br />

business performance and competitiveness, SMEs, due their scarcity of time, knowledge and resources,<br />

have experienced problems implementing and maintaining successful alliances.<br />

Considering a SME as defined in (EU 2009), SMEs represent 99 percent of non-financial businesses in<br />

US and EU and, in addition to their presence, play a key role creating employment: about two-thirds of<br />

total employment in the private sector; economic growth: contributing to more than half of the total valueadded<br />

created by businesses, innovation and social integration (Audretsch et al. 2009). This influence is<br />

not exclusive of western countries and works like (Zeng et al. 2010) present similar figures for Asian<br />

countries like China, Hong Kong or Taiwan.<br />

However, the above-mentioned studies also show that small firms have lower profitability, employee<br />

compensation, and labour productivity than large enterprises. In this context collaboration emerges as a<br />

potential solution but most of firms are reluctant to collaborate even if there is evidence of mutual benefits.<br />

In (Estanyol 2010), the author identifies a set of problems and barriers SME face in order to adopt<br />

collaborative approaches. Among of them, the self-centred behaviour of participants giving priority to their<br />

individual interests or for fear that others will take advantage of their co-operative decision is one of the<br />

most important. In addition, the lack of resources, mainly time, money and personnel, force SMEs to focus<br />

on their daily activities, losing the opportunity to create new business, enter new markets or create<br />

new products in collaboration with other SMEs.<br />

On the other hand, as summarised in (Nooteboom 1994) traditionally large corporations have been more<br />

efficient using external networks. They take advantage of their higher capital reserves and negotiation<br />

power to obtain favourable alliances with other organizations while SMEs need to find imaginative ways<br />

to survive. In uncertain economic periods, this gap increases and the ability to collaborate with others can<br />

make the difference between surviving or not. Fortunately, nowadays this situation is changing and the<br />

emergence of Web 2.0 technologies like DBE are creating a new range of options for collaboration, permitting<br />

SMEs to combine the use of these tools with their major flexibility in order to overcome their<br />

weaknesses by developing smart alliances and to reduce the gap with large corporations (Narula 2004).<br />

187


Francesc Estanyol<br />

The objectives of this work are as follows: a) to study the current context of SME collaboration in order to<br />

b) analyse how Web 2.0 technologies facilitate the emergence of co-operative behaviour and c) they role<br />

as enabler of business collaboration. Knowing the current situation permits to understand the problems<br />

SME face in order to adopt innovative ICT tools and to d) introduce DBEs as an alternative to facilitate<br />

businesses co-operation. Finally, the main objective of this study is to e) define and present MADBE, the<br />

first ecologically inspired multi-agent to DBE and its main contribution to research.<br />

2. The importance of Web 2.0 for co-operation<br />

During the last decade, ICT and Internet have changed our habits in a wide range of activities. Nowadays,<br />

it is difficult to understand our lives without e-mails, on-line tickets, digital photography or mp3 music.<br />

Although important, is not until the advent of Web 2.0 that the use of the Internet can be defined as<br />

revolutionary. Before that, services and content were made by IT specialists and users had a passive role<br />

interacting with them, simply viewing or downloading it. Software companies and Internet businesses<br />

were the players, the owners and who decided what, where, how and when to publish information on the<br />

World Wide Web (WWW).<br />

The term Web 2.0, firstly coined by Dancy DiNucci in 1999 but popularised by Tim O’Reilly in (O’Reilly<br />

2005) refers to the evolution of the Web and it is associated with applications where users are the main<br />

actors creating the content of the WWW. Although most researchers agree on some common concepts<br />

and principles around Web 2.0, currently there is no accepted definition. In this work, the author uses the<br />

term Web 2.0 as defined in (Hoegg et al. 2006), where the authors clearly separate the concept from<br />

specific technologies and define it as a philosophy or a new way to create web content which implies a<br />

greater collaboration among Internet users and the possibility to create communities.<br />

Independently of the scientific definition, most Web 2.0 applications like Facebook, Twitter, Flickr, Blogger<br />

or Wikipedia share some principles such as information sharing, interoperability, collaboration and cooperation.<br />

But the most powerful consequence of Web 2.0 is that it permits the creation and sharing of<br />

content in real-time. As pointed out by Clay Shirky in (Shirky 2009), nowadays it is possible to be informed<br />

about news, local traffic, deals or just friend information using these types of applications and all<br />

this faster than the official information channels such as TV, radio or newspapers.<br />

2.1 Self-organisation and collaborative behaviour<br />

With this revolution, the role of users has changed radically during the last years. They evolved from passive<br />

to active and nowadays are able to organise themselves, build communities and share information<br />

independently of the external channels.<br />

Web 2.0 facilitates self-organisation of participants and the emergence of collaborative behaviour, two<br />

key aspects to ensure co-operation success. Nowadays, individuals are able to identify something they<br />

would like to do, find others who would enable that action to be a success and access the resources to<br />

move to action. When users are involved in these types of actions, collaborative behaviour emerges naturally<br />

and facilitates the sharing of resources and the creation of a feeling of belonging to a community and<br />

therefore, makes users less dependent on traditional channels.<br />

Web 2.0 applications like CouchSurfing.com, HomeExchange.com or Compartir.org enable users to<br />

specify their interests (e.g. location, days, traffic route, timetable, etc.) and meet others offering the same.<br />

Apart from obvious economic and environmental benefits, these initiatives help to establish new relationships<br />

between users and to promote the creation of social communities with similar preferences.<br />

These tools have created a new way to perform old activities and represent an alternative to traditional<br />

services and products. This is possible because with Web 2.0 users are able to comment inputs, make<br />

recommendations and evaluate, rate, vote or tag users’ information. These actions permit to rank the<br />

users depending on their reputation and help the creation of trust and collective intelligence within the<br />

community.<br />

3. Web 2.0 as enabler for business collaboration<br />

The examples presented in previous section are useful to illustrate how Web 2.0 is changing people’s<br />

habits and routines but businesses, and SMEs in particular, are far from this situation. Despite of abundant<br />

proofs that the Internet is effective for gaining a competitive advantage, most small firms are not<br />

making full use of it (Martin and Matlay 2003).<br />

188


Francesc Estanyol<br />

Favoured by the declining costs of technology, businesses currently have a wide range of options to organise<br />

themselves and improve the relationships with their providers/clients. Business networking applications<br />

like LinkedIn; Business Intelligence (BI) tools like Pentaho or MicroStrategy; or Customer relationship<br />

management (CRM) solutions permit to control organisational activity more effectively, support<br />

the decision making process, establish links with other partners or facilitate co-operation processes that<br />

require flexibility and trust (Caby-Guillet et al. 2006).<br />

But compare with individuals, most firms are still in the first stage of the Internet revolution and are using<br />

these tools mostly for upgrading existing mechanisms, for example, replacing postal mail by e-mail or<br />

communicating their corporate vision using websites instead of other traditional media but not for creating<br />

new ways of doing businesses or interact with other SMEs, missing the real possibilities that Web 2.0<br />

brings (Hoegg et al. 2006).<br />

In this context, recently DBEs have emerged as a new Web 2.0 paradigm to solve some of these problems.<br />

DBEs permit to access dynamically to valuable information about other participants (e.g. background,<br />

previous collaboration, results, etc.) and to introduce reputation, voting or tagging systems to<br />

ensure security, trust and commitment among the participants. These characteristics promote the creation<br />

of digital environments in which businesses co-evolve and collaborate, permitting enterprises to<br />

adapt quickly to a changeable environment and to save resources.<br />

The next section describes DBEs and how SMEs can benefit from the opportunities that this new paradigm<br />

offers.<br />

3.1 Digital business ecosystems<br />

Although the definition of a DBE may differ considerably depending on the author, in this work it refers to<br />

those digital environments that reproduce natural ecosystems and in which businesses can interact in a<br />

similar way to what people are doing in social networks.<br />

DBEs permit the interoperability of firms thorough the implementation of intelligent mechanisms of publishing<br />

characteristics, services and needs. In this way, SMEs are able to find other partners and to establish<br />

potential relationships with other entities of the ecosystem in a similar way individuals are doing in<br />

the applications presented in Section 2.1.<br />

Moreover, they are an innovative approach to supporting the adoption and development of ICT, which, as<br />

identified in (Nachira 2004), is a crucial element for improving productivity and for enabling business networking<br />

in a competitive knowledge-based economy.<br />

Like other Web 2.0 tools, DBEs enable the self-organisation of their participants that in the case of SMEs,<br />

can lead to improved business networking, greater competitiveness and a better exploitation of new opportunities.<br />

In these environments, businesses are represented by software components that exchange information,<br />

create coalitions and find new opportunities to compete and improve their individual position within the<br />

digital ecosystem but also for the benefit of the community they belong to. In this scenario, SMEs live,<br />

interact, evolve and become extinct as natural species do in natural ecosystems, and working together,<br />

create collaborative behaviour that leads to an improvement of the competitiveness of the whole set of<br />

cooperating companies.<br />

In order to be considered a digital ecosystem, the collection of autonomous digital entities must present<br />

the following characteristics:<br />

Complexity. According to (Lewin 2000), a complex system is one whose properties are not fully explained<br />

by an understanding of its parts.<br />

Self-organisation. Self-organisation is defined as the ability of complex systems to create new order<br />

and coherence. It relies on the spontaneous emergence of order by the interaction of individuals in an<br />

ecosystem.<br />

Emergence. Emergence is a concept closely linked to self-organisation and evolution and it is defined<br />

as the process which creates new order together with self-organisation.<br />

189


Francesc Estanyol<br />

Co-evolution. Defined in (Mitleton-Kelly 2003) as the evolution of one domain or entity is partially<br />

dependent on the evolution of other related domains or entities, it means that one domain or entity<br />

changes in the context of the others.<br />

Adaptation. Adaptation is an old concept from Darwin’s “Origin of Species” and it refers to the capacity<br />

of individuals to adapt to a changing environment.<br />

In addition to these principles, a DBE must be loosely coupled, meaning that each participant does not<br />

need to expose the details of its internal work-flows and requires only that it presents a service interface.<br />

Although DBE is a relatively new concept, recent projects like (OPAALS 2007) or (ONE 2006) have<br />

shown its potential to promote collaboration. These initiatives focused on providing organisations in the<br />

service industry with more sophisticated mechanisms to concentrate and publish services, improve negotiation<br />

and facilitate learning.<br />

Nonetheless, most of these tools use the Service Oriented Architecture (SOA) paradigm that, although it<br />

has proved to be successful in works like (Briscoe and De Wilde 2006), do not permit to implement all the<br />

characteristics of DBEs, limiting their use and potential. In that sense, Multi-Agent Systems (MAS)<br />

emerge as a perfect alternative because fulfil all DBEs principles. Moreover, MAS have been applied<br />

successfully in a large number of disciplines, from transport systems like (Balbo and Pinson 2005) to<br />

medical automatic diagnosis systems like HealthAgents (Lluch-Ariet et al. 2008).<br />

By definition, MAS facilitate the implementation of distributed architectures in which each entity is represented<br />

by an autonomous intelligent agent which pursues its goals and executes its tasks trying to improve<br />

its performance considering the conditions of the environment and interacting with other agents.<br />

Despite the parallelism between both paradigms seems clear, it is difficult to find MAS approaches to<br />

DBE and currently there is no application focused on SMEs. The work presented in (Muntaner-Perich and<br />

De la Rosa Esteva 2007) and the EveSim simulation tool resulting from the OPAALS project are the only<br />

initiatives to have succeeded in the creation of a multi-agent tool for DBEs but they failed on implementing<br />

the concepts of natural ecosystems.<br />

4. MADBE, a multi-agent digital business ecosystem for SME collaboration<br />

In moving beyond the current approaches and understanding digital ecosystems from an artificial intelligence<br />

perspective, MADBE represents the first ecologically inspired multi-agent digital business ecosystem<br />

for SME collaboration. In MADBE, the digital entities (SMEs) are represented by agents with their<br />

characteristics, services and needs. Therefore, the interactions with other agents are the representation<br />

of business-business relationships.<br />

Apart of the new technological approach, the step forward from similar initiatives is the introduction of<br />

mutualistic ecological network concepts to define the business interactions. Digital entities are defined by<br />

rules found in natural ecosystems in order to reproduce the characteristics of mutualistic networks, where<br />

species interact in a co-operative way and benefiting each other. Additionally, as summarised in (Bastolla<br />

et al. 2009), this type of network minimises competition and promotes species diversity and stability to the<br />

network itself, resulting in an environment with a larger number of SMEs collaborating with each other<br />

and establishing long-term mutual benefit relationships.<br />

Following these principles, the agents in MADBE are defined based on the set of ecological concepts<br />

presented in next section. This set of features are a starting point for designing the particular characteristics<br />

of each of the entities willing to join the digital ecosystem and create the necessity, covered by this<br />

study, to translate them into business concepts defining SME collaboration. In order to define the business<br />

concepts to include in MADBE regarding SME collaboration, the author used the findings of (Estanyol<br />

2010), a study which summarises why SMEs should co-operate, the problems they face adopting<br />

collaborative approaches and the factors influencing interfirm collaboration effectiveness. Therefore, the<br />

mapping process presented in next section is based on the results of that work, which permitted to understand<br />

SME collaboration and to select only the relevant business concepts, discarding ecological concepts<br />

with no applicability in the business domain and vice versa.<br />

190


Francesc Estanyol<br />

4.1 Mapping ecological concepts to the domain of SME collaboration<br />

The combination of natural ecosystems and the characteristics of real SME collaboration result in the<br />

definition of the MADBE agents by the following ecological concepts:<br />

Traits: a set of characteristics that define any entity within the system. In terms of SMEs, these traits are<br />

related to the characteristics of the firm and its expertise/interest in collaboration. The following items<br />

were selected in order to find complementarities and compatibilities between SMEs and facilitate partner<br />

search and selection. They cover most of the factors identified to do with collaboration success and to<br />

overcome the barriers to collaboration identified in (Estanyol 2010).<br />

Number of employees.<br />

Age of the company.<br />

Type of SME (individual enterprise, SL, SA, association, other).<br />

Goal. A short description of the firm’s global vision and main objectives.<br />

Average employee education level (no studies, undergraduate, graduate, postgraduate).<br />

Average age of employees (less than 25, between 25 and 35, between 35 and 50, more than 50).<br />

IT skilled personnel (yes/no).<br />

Technological capability (none, low, medium, high).<br />

Management style (owners, owners with management studies, professionals).<br />

Interest in co-operation (no interest, low, medium, high).<br />

Previous collaborations (yes, number of collaborations/no).<br />

Previous collaboration with the candidate (yes/no).<br />

Currently involved in co-operation projects (yes, number of projects/no).<br />

Level of satisfaction with previous collaborations (unsatisfied, low, medium, high, very high).<br />

List of services and resources (core competences). An array of a dynamic list of services and resources.<br />

The company has the option of adding non-existing items, making them available to other<br />

firms and increasing the list of system known services and resources.<br />

List of needs and resources. The same concept of the previous traits but about the needs of the<br />

firms.<br />

Degree of complementarity: the degree to which any trait of a given info-habitant is complementary to<br />

another trait possessed by another info-habitant. In MADBE, for each service, a degree of complementarity<br />

is given. It is normally a value between 1 and 5. For example, if a SME is looking for local transport, it<br />

could specify a maximum distance from its location. In this case, the degree of complementarity depends<br />

on the distance from the candidate to this location, the closest being the maximum. Although this ranking<br />

method has been adopted as a standard approach to define the degree of complementarity, the possibility<br />

of specifying different types of services and resources implies to treat each case individually and to<br />

create a system which permits the users to define how they want to specify the complementarity of their<br />

core competences.<br />

Habitat and Meta-communities: habitat refers to the place where an organism most commonly occurs.<br />

Basic market information like trend or the type of industry items is included in this section. Additionally,<br />

MADBE also needs to include those characteristics of the environment which influence the activity of the<br />

SME. In that sense, as it has been identified that SMEs look for external help and public funding when it<br />

comes to collaboration, MADBE includes a couple of lists indicating the presence of collaboration agents<br />

and funding organisations in the environment. In addition, SMEs tend to collaborate with firms from the<br />

same area due their preference for face-to-face contact, which required the inclusion of the firm location.<br />

On the other hand, meta-communities are aggregations of entities that belong to different habitats (or<br />

regions) and that occasionally interact when certain conditions are met. This concept is important for<br />

those situations when SMEs are looking to collaborate within the framework of a bigger association, not<br />

individually.<br />

Type of industry (monopoly, oligopoly, perfect competition).<br />

Market trend (unknown, downward, stable, upward).<br />

191


Francesc Estanyol<br />

Government regulations (low, medium, strict, very strict).<br />

List of collaboration agents. Information about agents which can help and provide support to SMEs<br />

during the collaboration process.<br />

List of public/private funding organisations. List of organisations promoting and funding collaborative<br />

initiatives.<br />

SME Location.<br />

A list containing the associations which the SME is a member (cluster, science park, etc).<br />

Niche: the niche is the “job” of the organism within the habitat. This concept groups those characteristics<br />

about the interactions and role of the agent in the ecosystem. The niche concept is also useful in explaining<br />

how seemingly similar species can coexist in the same community. In MADBE, the niche is the multidimensional<br />

space composed by the identified characteristics of the habitat and the way agents interact<br />

with their environment. Therefore, it is necessary to include the business context of the company and its<br />

sector. Additionally, SMEs within small areas tend to have similar world-views and to speak similar languages,<br />

which have been identified as key factors for collaboration success. It is for those reasons that,<br />

apart from location and culture, it is important to include the languages known by the company. Finally,<br />

the niche concept also explains how species from a specific habitat collaborate temporarily with species<br />

from a different environment, which is why it has been included a parameter to determine the willingness<br />

of the company to create new businesses in different business contexts.<br />

Business context (local, national, <strong>European</strong>, international).<br />

Business sector (research, innovation, product development, marketing, etc.)<br />

Interest in creating new businesses in a different business domain (low, moderate, high, very high).<br />

Languages. List of languages known by the company.<br />

Size of the SME in relation to the market size (very small, small, medium, big, very big).<br />

Employee mobility (yes/no). Parameter to detect the capacity of the company to send employees to<br />

other companies for learning and transfer of knowledge purposes.<br />

Fitness: a measure that is used to determine how well an entity is doing during its lifetime in the digital<br />

ecosystem. An intelligent tool like MADBE needs fitness measures in order to evaluate the different options<br />

during simulations and to provide the optimum results to the user. Unfortunately, there is no consensus<br />

about a specific set of measures to determine the fitness of SME collaboration. Some authors<br />

defend objective measures like the duration of the co-operation or the number of sales while others prefer<br />

to consider the subjective perception of their participants. MADBE has two set of parameters in order to<br />

implement both. A first group including basic accounting concepts and useful to evaluate the fitness of<br />

the SMEs within the ecosystem and a second group including those parameters to evaluate the collaboration<br />

process itself.<br />

Individual SME performance:<br />

Number of current collaborations. Useful to identify the role and the influence of the company in the<br />

ecosystem.<br />

Average of previous collaborations duration.<br />

5 year annual turnover (i.e. sales).<br />

5 year proportion of annual turnover of new products/services.<br />

5 profit margin = pre tax net profit/turnover.<br />

5 year RoA (return on assets) ratio.<br />

5 year Liquid Ratio = (current assets - inventories) / current liabilities.<br />

5 years percentage increase of number of employees.<br />

Performance of the collaboration:<br />

Reputation rate. An array of evaluations including the opinion about the SME from partners with<br />

whom it has collaborated. A reputation mechanism to overcome the trust, credibility and compromise<br />

problems and to evaluate the co-operation skills of the company. This measure permits to identify reliable<br />

partners and discard partners that break collaborations or fail to carry out their assigned tasks.<br />

Duration of the co-operation.<br />

192


Francesc Estanyol<br />

Number of new products/services. The proportion of current sales from products newly introduced<br />

during the co-operation.<br />

Subjective perception (bad, neutral, positive, very positive).<br />

Number of new jobs created.<br />

Access to new markets (yes/no).<br />

New relationships (yes/no).<br />

The concepts of traits, niche and fitness have been employed for the design of the agents interacting and<br />

living together within MADBE, and are therefore responsible for much of the behaviour displayed by them<br />

in any given interaction. More details about the ecological concepts and their role specifying rules of interactions<br />

and specific details of MADBE architecture can be found in (Lurgi and Estanyol 2010).<br />

5. Conclusions<br />

On the basis of the findings from the literature review, the author explored the current context of SME<br />

collaboration and how Web 2.0 applications are changing the way individuals interact. By using current<br />

examples, it has been demonstrated that nowadays users are able to self-organise themselves creating<br />

communities in order to satisfy their needs independently of the traditional and external channels. These<br />

communities facilitate the emergence of collaborative behaviour among the participants but businesses,<br />

and in particular SMEs, are not taking profit of it.<br />

In order to understand why, the author analysed the type of existing ICT tools, concluding that most applications<br />

are not specifically designed considering SMEs needs and they focus on how improve their<br />

current business practices instead of looking for mechanisms to create new ways of doing businesses.<br />

In that context, the new paradigm of DBEs has been introduced as a possible solution to overcome these<br />

problems and, after studying the existing alternatives, the work concluded with the presentation of<br />

MADBE, the main contribution to research of this study and the first ecologically inspired multi-agent<br />

DBE.<br />

The author believes that DBEs can play an important role enabling and improving SMEs co-operation. In<br />

line with Nachira, who in (Nachira 2002) proclaims that “DBEs produce an extraordinary competitive advantage<br />

for SMEs”, the author thinks that DBEs designed and developed considering SMEs needs could<br />

benefit them in the long-term and enhance their chances of competing with larger organisations.<br />

In addition, the adoption of multi-agent technology in MADBE and the mapping between business and<br />

ecological concepts found in natural mutualistic networks permits to define an application fulfilling SME<br />

needs and truly inspired in ecology. Moreover, by treating SMEs as autonomous intelligent agents, the<br />

system requires a minimum level of dedication and permits to analyse the individual behaviour of each<br />

company in a collaborative environment.<br />

At this point, it is important to remark that this work only covered the definition of the ecological concepts<br />

from a business perspective, which is a fundamental step to demonstrate that the concepts explaining<br />

natural mutualistic networks of interaction could be translated into the business domain and report benefits<br />

for SMEs but the system is still under development and therefore, it has not been deployed to a real<br />

market scenario and there are no results yet. It will take time to develop the system and convince SMEs<br />

to use it in order to collect real data. In that sense, the current participation of the author in (EcoBusiness<br />

2010) ensures the continuity of this research.<br />

Another aspect to consider when evaluating DBEs as enablers for collaboration is that the utilisation of<br />

any Web 2.0 tool is not enough to ensure collaborative success. In order to become part of Web 2.0 it is<br />

necessary a change of mentality. E-business is more than e-commerce and it is not enough to use Web<br />

2.0 tools for collaboration without a planned strategy. It is obvious that companies cannot take the same<br />

risks as individuals and that their bureaucracy and risk aversion make sudden changes difficult, but it is<br />

also true that Web 2.0 opens up a wide range of business opportunities for collaboration and cooperation<br />

that companies should consider. In order to succeed, it is necessary that firms interested in<br />

collaboration have a clear understanding of why they want to co-operate and which results they expect<br />

before the process starts. Collaboration requires a favourable attitude and an analysis of the individual<br />

characteristics of the participants in order to find complementarities and synergies while avoiding the<br />

emergence of conflict.<br />

193


Francesc Estanyol<br />

Finally, a positive consequence of DBEs is that business environments are represented as complex systems,<br />

permitting to analyse and understand the principles of their formation, evolution and interrelationships<br />

and exploit the research carried out in other sciences like biology, mathematics, computer science<br />

or social science in the business domain. For example, by applying network analysis algorithms in a system<br />

like MADBE would be possible to identify the key firms of the ecosystem, the “leadership companies”<br />

or the ones who have a strong influence on the others. In that sense, MADBE will not benefit only the<br />

involved SMEs but also third parties like public governments, who, for example, could use it to identify<br />

which companies should receive more funding or those reaching undesired monopolistic positions.<br />

Acknowledgments<br />

The author wants to thank Miguel Lurgi for the specification of MADBE from an ecological perspective<br />

and for the identification of the key ecological concepts to be included in the system. This work is part of<br />

the EcoBusiness project, funded by the <strong>European</strong> Commission through the Seventh Frame work programme<br />

Marie Curie Actions – Industry-Academia Partnerships and Pathways (IAPP). Grant agreement<br />

no.: 230618.<br />

References<br />

Audretsch, D, van der Horst, R, Kwaak, T, and Thurik, R.. (2009) “First section of the annual report on EU small and<br />

medium-sized enterprises”, January.<br />

Axelrod, R. (1985) “The importance of being nice, retaliatory, forgiving and clear”, The Economist, pp 99–102, November,<br />

9.<br />

Balbo, F and Pinson, S. (2005) “Dynamic modelling of a disturbance in a multi-agent system for traffic regulation”,<br />

Decision Support Systems, Elsevier, Vol 41, No. 1, pp131–146.<br />

Bastolla, U, Fortuna, M, Pascual-García, A, Ferrera, A, Luque, B, and Bascompte, J. (2009) “The architecture of<br />

mutualistic networks minimizes competition and increases biodiversity”, Nature, Nature Publishing Group, Vol<br />

458, No. 7241, pp 1018–1020.<br />

Briscoe, G and De Wilde, P. (2006) “Digital ecosystems: Evolving service-orientated architectures”, Bio-Inspired<br />

Models of Network Information and Computing Systems. Vol 1, pp 1–6.<br />

Caby-Guillet, L, Crave, S, and Ladame, S. (2006) “How ICT can facilitate trust inside networks of SME: the role of<br />

professional virtual communities”, International conference on Concurrent Engineering, Milano<br />

Coombs, R, Richards, A, Saviotti, P, and Walsh, V. (1996) Technological Collaboration: The dynamics of cooperation<br />

in industrial innovation, Edward Elgar Pub, ISBN: 1858982359.<br />

Dickson, K. (1996) “How informal can you be? trust and reciprocity within co-operative and collaborative relationships”,<br />

International Journal of Technology Management, 11, Inderscience Publishers, Vol 1, No. 2, pp 129–<br />

139.<br />

EcoBusiness Website (2010) URL http://www.ecobusiness.cat, Accessed date 19/11/2010.<br />

Estanyol, F. (2010) “The SME co-operation framework: A multi-method secondary research approach to SME collaboration”,<br />

Proceedings of ICEME 2010, Hong Kong.<br />

EU. (2009) “On the implementation of commission recommendation of 6 may 2003 concerning the definition of micro,<br />

small and medium-sized enterprises”, Commission Staff Document, October.<br />

Hoegg, R, Meckel, M, Stanoevska-Slabeva, K, and Martignoni, R. (2006) “Overview of business models for web 2.0<br />

communities”, proceedings of GeNeMe, pp 23–37.<br />

Hoffmann, W, H and Schlosser, R. (2001) “Success factors of strategic alliances in small and medium-sized enterprises<br />

- An empirical survey”, Long Range Planning, Elsevier, Vol 34, No. 3, pp 357–381.<br />

Lewin, R. (2000) “Complexity: Life at the edge of chaos”. University of Chicago Press.<br />

Lluch-Ariet, M, Estanyol, F, Mier, M, Delgado, C, González-Vélez, H, Dalmas, T, Robles, M, Sáez, C, Vicente, J and<br />

Huffe, Sl. (2008) “On the implementation of healthagents: Agent-based brain tumour diagnosis”, Agent Technology<br />

and e-Health, pp 5–24.<br />

Martin L. M., and Matlay, H. (2003) “Innovative use of the internet in established small firms: the impact of knowledge<br />

management and organisational learning in accessing new opportunities”, Qualitative Market Research: An International<br />

Journal, MCB UP Ltd, Vol. 6, No. 1, pp 18 – 26.<br />

Mitleton-Kelly, E. (2003) “Ten principles of complexity and enabling infrastructures”, Complex systems and evolutionary<br />

perspectives on organisations: The application of complexity theory to organisations, Pergamon, pp 23–50.<br />

Muntaner-Perich, E and de la Rosa Esteva, J. (2007) “Using dynamic electronic institutions to enable digital business<br />

ecosystems”, Coordination, Organizations, Institutions, and Norms in Agent Systems II, Springer, pp 259–273.<br />

Nachira, F. (2002) “Towards a network of digital business ecosystems fostering the local development”. URI<br />

http://hdl.handle.net/2038/529<br />

Narula, R. (2004) “R&D collaboration by SMEs: New opportunities and limitations in the face of globalisation”, Technovation,<br />

Elsevier, Vol 24, No. 2, pp 153-161.<br />

Nooteboom, B. (1994) “Innovation and diffusion in small firms: theory and evidence”, Small Business Economics,<br />

Springer, Vol 6, No. 5, pp 327–347.<br />

O’Reily, T. (2005) “Web 2.0: Compact definition?” Blog, October, URL http://radar.<br />

oreilly.com/archives/2005/10/web-20-compact-definition.html, accessed date 16/06/2010.<br />

ONE. (2006) One negotiation environment project. Website, URL http://one-project.eu/<br />

site/modules/content/index.php?id=1, accessed date 08/08/2010.<br />

194


Francesc Estanyol<br />

OPALS. (2007) Open philosophies for associative autopoietic digital ecosystems. Website, URL http:<br />

//www.opaals.eu/, accessed date 15/08/2010.<br />

Shirky., C (2009) “How social media can make history”, <strong>Conference</strong> Talk, TED@State, June, URL<br />

http://www.ted.com/talks/clay_shirky_how_cellphones_twitter_facebook_can_make_history.html, Accessed<br />

date 03/07/2010.<br />

Varamaki, Eand Vesalainen, J. (2003) “Modelling different types of multilateral co-operation between SMEs”, Entrepreneurship<br />

& Regional Development, 15:2747.<br />

Zeng, S, Xie, X, and Tam, C. (2010) “Relationship between cooperation networks and innovation performance of<br />

SMEs”, Technovation, Elsevier, Vol 30, No. 3, pp 181 – 194.<br />

195


Ideas About Profitability in Research and Development and<br />

the Selective Pressure From Management Accounting<br />

Albrecht Fritzsche<br />

Technical University, Darmstadt, Germany<br />

a.m.fritzsche@gmx.net<br />

Abstract: The selective pressure on the development of products and production technologies is not exerted directly<br />

by customer demand, but by its interpretation in a company. The decisions on technical development usually take place<br />

a long time before a product reaches the market on the basis of calculations about profitability in management accounting.<br />

The profitability can depend on various factors, including production facilities, supply networks, intellectual<br />

property, distribution channels and competition. The calculations that lead to decisions on an investment tend to be<br />

very complex. Nevertheless, they do not represent fully realistic market conditions, but only a simplification for the sake<br />

of reliable estimates. In a conventional simple evolutionary model of innovation, technical development does not express<br />

expert knowledge. In addition, it adapts as much to the shortcomings of the financial calculations as to the<br />

prognostic information they are based on. If the selective pressure from management accounting is high, technical<br />

development is likely to focus on the exploitation of the missing cost factors in the calculations and neglect the reality of<br />

market demand. In practice, innovations in research and development are not propelled by arbitrary change. Expanding<br />

the conventional model, the expertise in research and development can be represented by introducing intentional<br />

improvement operators. These operators show a potential to avoid unwanted directions of adaptation, which<br />

gives reason to believe that common sense in research and development play an important role in avoiding detrimental<br />

effects of incomplete cost calculations.<br />

Keywords: computational models of innovation, investment decisions on R&D, technology assessment and management<br />

accounting, simulation of evolutionary adaptation, profitability calculations and human expertise<br />

1. Background<br />

During the last decades, it has become very popular to use Genetic Algorithms for the solution of complex<br />

computational problems. In contrast to conventional optimization techniques, such as gradient-based or<br />

simplex-based methods, Genetic Algorithms require only a minimum of analytic representation of the<br />

problem structure. They can therefore also be applied to problems with a very high level of complexity.<br />

Genetic Algorithms represent a heuristic approach to problem solving. There is no guarantee that they<br />

produce optimal results. Nevertheless, they have proven to be very useful in a large variety of different<br />

scenarios, from factory scheduling over jet engine design to financial portfolio management (see Goldberg<br />

2002).<br />

Thanks to their success as a search method, Genetic Algorithms have become the object of extensive<br />

scientific research. The concept of Genetic Algorithms is based on the principle of evolutionary adaptation.<br />

Evolutionary approaches are quite common in many different disciplines, for example in economics, sociology<br />

and philosophy (Dennett 1996, Broesel et al. 2007). In economy, evolutionary models are used, for<br />

example, to describe the dynamics of technological change. While analytic approaches focus on the<br />

identification of optimal economic transactions and the way how they can be achieved, evolutionary approaches<br />

put the emphasis on the explanation of innovative processes in complex environments where<br />

optimal, equilibristic states are unknown (Dopfer 2008).<br />

2. Evolutionary approaches to technical development<br />

2.1 Technology, economics and evolution<br />

Evolutionary approaches in economy are strongly related to the idea of a dynamic relationship between<br />

supply and demand on a free market. Consumption and production are interdependent: Consumers react<br />

to price changes and the introduction of new products; producers offer goods up to the extent to which<br />

selling them is profitable. Technical development, in terms of new products as well as new production<br />

methods, is an adaptive process that can be compared to natural evolution, which, according to Darwin,<br />

advances through repetitive arbitrary changes under selective pressure from outside (see e.g. Hodgson<br />

1993). Finding exact economic correspondences for the reproductive processes in nature, the genetic<br />

carriers of information and the genotype of adaptation is rather difficult. The general principle of evolution<br />

as an adaptive activity without intentional direction, however, applies very well to the dynamics business on<br />

an unrestricted market (see e.g. Broesel t al. 2007, Beck-Sickinger/ Petzold 2009).<br />

196


Albrecht Fritzsche<br />

The idea to describe the development of trade and industry similar to the biological models of natural<br />

adaptation and selection dates back to the late nineteenth century (Marshall 1890, Veblen 1898). In the<br />

twentieth century, Joseph Schumpeter has been very influential with his evolutionary approach to innovation<br />

(Schumpeter 1912). Schumpeter describes innovation as a source of continuous change, driven by the<br />

search for competitive advantage of the companies on a market. After several decades in which economic<br />

research showed little interest in evolutionary models, the approach regained increasing popularity during<br />

the 1980ies (Hodgson 2005), mainly due to the seminal paper by Richard Nelson and Sidney Winter<br />

(Nelson/ Winter 1982).<br />

Evolutionary economics is a relatively open field of research, which has been influenced by various scientific<br />

concepts, such as bounded rationality or complexity studies (Simon 2005, Frenken 2006). It is concerned<br />

with both microscopic and macroscopic phenomena of economic adaptation and their dependency<br />

on socio-cultural or administrative structures. It is characteristic for evolutionary economics to assume that<br />

the economic development cannot be guided and controlled (comp. Dopfer 2008, Winter 2005). Nevertheless,<br />

it is possible to influence the process of their adoption on a market, which depends on their representation,<br />

distribution and the selective pressure exerted on them.<br />

2.2 Simulated evolution in Genetic Algorithms<br />

An important source of influence on the studies of evolutionary processes in general has been the development<br />

of electronic data processing and the rise of modern information technology. Computers can<br />

execute a large number of logical operations in short time; they are therefore a very useful tool to simulate<br />

adaptive phenomena that are caused by the combination of many simple change routines. It also turned<br />

out very quickly that it was possible to apply these simulations of evolutionary developments and similar<br />

procedures in solving optimization problems in complex search spaces. For this purpose, the core elements<br />

of the evolutionary process were formalised in the concept of what is now known as Genetic Algorithms.<br />

Figure 1 shows a rough sketch of the design of Genetic Algorithms (see e.g. Goldberg 1998). The search<br />

starts with a set of possible alternatives to solve the problem, called a population. In course of the search,<br />

the alternatives in the population repetitively go through some changes, usually caused by mutations or<br />

re-combinations of the alternatives.<br />

Figure 1: Concept of evolutionary search in Genetic Algorithms<br />

197


Albrecht Fritzsche<br />

As the genetic metaphor indicates, these operations are inspired by processes in biological reproduction.<br />

With a mutation, certain characteristics of the alternatives are modified at random; a re-combination puts<br />

together arbitrary characteristics from two (parent-) alternatives in order to create a new one. The operations<br />

have no knowledge about the quality of the alternatives. The evaluation of the changes in the population<br />

happens separately in a subsequent step of the procedure, where the best alternatives are selected<br />

to form a new population. Since these changes are performed on alternatives which already possess<br />

positive characteristics, it can be expected that the population will step by step develop into set of optimal<br />

alternatives, similar to the survival of the fittest in nature.<br />

2.3 Imitation and innovation<br />

Genetic Algorithms are just one heuristic search method among many others, but they play an outstanding<br />

role as a representation of evolutionary processes. In fact, Genetic Algorithms have widely replaced the<br />

biological model of natural development as the point of reference for the concept of evolution, thanks to the<br />

simplifications they provide and their implementation on computers. It is much easier to understand the<br />

principle of evolution on the example of a Genetic Algorithm than by looking at the development of species<br />

in nature, which retains importance as the origin and the symbol for the general importance of evolutionary<br />

processes.<br />

In particular, Genetic Algorithms have been used to research how technical change becomes possible.<br />

According to David Goldberg, the basic reproduction methods of Genetic Algorithms, the mutation and the<br />

re-combination, in combination with the selective pressure from outside, represent the two fundamental<br />

activities in research and development (Goldberg 1998, Goldberg 2002). The combination of mutation and<br />

selection describes continuous improvement processes as they are addressed, for example, by the kaizen<br />

principle. This improvement leads to a higher efficiency of the processes that are already implemented. In<br />

that sense, Goldberg understands it as imitation. The combination of selection and re-combination, on the<br />

other hand, leads to improvement by a new arrangement of the single elements that define a technical<br />

solution. The new arrangement changes the effect of technology rather than its efficiency. Inasmuch as<br />

technology is defined by its effect, the result of the change can be considered as an innovation.<br />

Topologically, mutations and re-combinations describe two different movements in the search space.<br />

Mutations cause arbitrary small changes. Most of the original element of the search space remains intact.<br />

Re-combinations, on the other hand, bring the characteristics of two different elements together. If such a<br />

change had to be generated solely by mutations, it would require a huge number of steps. Considering the<br />

selective pressure under which the changes are taking place, it is highly improbable that mutations lead to<br />

similar results as re-combinations. At the same time, re-combinations are not very effective in generating<br />

small changes of one original alternative.<br />

In general, it is hardly possible to identify the best elements of the search space, if mutations or<br />

re-combinations are used exclusively (see Goldberg 2002). Both of them have to be used together. For<br />

some problems, it may be necessary to apply re-combinations more frequently than mutations; for other<br />

problems, the opposite may be true. In addition, the size of the population is also an important parameter<br />

for the success of Genetic Algorithms as optimization tools. Other parameters have also been researched<br />

(Reed et al. 2001). Genetic Algorithms therefore are no "free lunch" (Wolpert/ MacReady 1997); they do not<br />

provide a generally applicable solution technique. Instead, they have to be adapted to the given problem<br />

situation. In Goldberg's model, this can be understood as the need for innovation management.<br />

3. The influence of accounting on technical development<br />

3.1 The paradox of heuristics and analytics<br />

The strength of the evolutionary approach in the different fields of research is mainly due to the fact that it<br />

shows how improvement is possible without knowledge of the reasons for preference (see e.g.<br />

Beck-Sickinger/ Petzold 2009). It does not matter why certain alternatives are considered to be better than<br />

others. It is only necessary to know that they are better in order to initiate the selection procedure. Apart<br />

from the current set of alternatives and the operators to identify new alternatives, the evolutionary process<br />

does not carry any further information. Economy, on the other hand, is the expression of a systematic activity<br />

of understanding and enabling improvement. The idea of stable economic structures that do not<br />

evolve themselves over time is contradictory to the evolutionary approach. Economic structures can be<br />

considered as expressions of expertise in improvement, both in a deductive and an inductive way. The<br />

notion of quantifiable goods and their prices represents, for example, certain ideas about what can be<br />

198


Albrecht Fritzsche<br />

improved; the notion of trade and production represents certain ideas about the way how this improvement<br />

takes place.<br />

The strength of the evolutionary approach lies in the fact that it does not require any expertise to explain<br />

improvement. Quite in the contrary, it shows how arbitrary change can lead to improvement under selective<br />

pressure. Research in evolutionary economics is therefore confronted with one fundamental question: How<br />

do evolutionary adaptation and economic expertise connect? The following pages give an example of such<br />

a connection. They also show that research on Genetic Algorithms and similar search methods can help to<br />

identify various different levels on which evolutionary adaptation and economic expertise come together.<br />

Our study explores the relation between departments for management accounting and for research &<br />

development. Both departments have a strong influence on technical change, based on certain systematic<br />

activities to enable and assess improvement. At first sight, it might seem that the installation of these departments<br />

contradicts the understanding of technical development as an evolutionary process. This paper<br />

shows, however, that there is good reason to approach the interaction of these departments in evolutionary<br />

terms. In particular, the paper uncovers certain parallels between technical development as an evolutionary<br />

process and recent experiences with the application of Genetic Algorithms for optimization in complex<br />

systems.<br />

3.2 The split between entrepreneur and engineer<br />

The high frequency in which companies introduce new products to a market makes it easy to forget that<br />

technical innovation is a slow process. Cars, computers, cell phones or medicine require a long period of<br />

preparation, both for research and development of the products themselves and for the establishment of<br />

production facilities, skills and supply networks. The investments during this phase are considerable. The<br />

most important decisions about technology therefore take place a long time before it enters a market. At<br />

this time, however, there are no sales figures yet that could exert a selective pressure. Instead, the decisions<br />

are made on the basis of estimates about the expected profitability as marketing and accounting<br />

departments are able to supply them (see e.g. Trott/ Lataste 2005).<br />

For the sake of a brief illustration, the difference between the reference to the market and the reference to<br />

investment calculations in accounting can be, very roughly, described in the following way: It starts with<br />

naming the dimensions in which the profitability of the innovation will be measured. This includes, for<br />

example, the raw material and parts supply, labour, energy, maintenance and services as well as the<br />

one-time effort to implement the production facilities. In these dimensions, the innovation can be expressed<br />

as a transformation vector that shows the ratio of the quantity of the goods consumed in production to the<br />

quantity of the good produced. Marketing research then provides estimates for customer demand and the<br />

prices of the different goods. Knowing the implementation effort, the transformation vector of production<br />

and the prices, it becomes possible to calculate the expected profitability of the innovation. If the evaluation<br />

of the innovation referred directly to the market, it would not be necessary to establish these dimensions.<br />

The profit could simply be measured by looking at the bank account.<br />

As a consequence of replacing the market response as the selective pressure from outside by investment<br />

calculations, the process of technical development is split in two. In the first part of the process, market<br />

response in approximated in the dimensions of the calculation and the pricing vector. This approximation is<br />

subject to constant change, because it is impossible to consider all sources of influence on the market. It<br />

can therefore be understood as an evolutionary process by itself. Since an entrepreneur is essentially<br />

somebody who discovers new ways to make profit, the approximation of market response for investment<br />

calculations reflects the entrepreneurial approach to innovation. In the second part of the process, research<br />

and development activities change technology under the selective pressure from accounting. Inasmuch as<br />

accounting measures profitability in certain dimensions, this process resembles a multi-objective optimization<br />

problem, with slowly changing objectives over time (see Table 1).<br />

3.3 The complexity problem of profitability calculations<br />

Considering the multiple dimensions of profitability, it is clear that the improvement of a product or a production<br />

process with respect to material costs, product utility or any other single objective does not necessarily<br />

lead to an overall better solution. Quite in the contrary, an improvement with respect to one objective<br />

is most likely to have detrimental effects with respect to other objectives. The investment calculations<br />

in accounting can display these effects, but it is quite a different question whether it is also possible to<br />

predict them. Some directions of research and development may sound very promising in advance, but<br />

199


Albrecht Fritzsche<br />

later turn out to be to expensive, considering all factors involved. This phenomenon is, for example, widely<br />

known in supply chain management. It can be described as "over-complexity" (Klaus 2005).<br />

Table 1: Appearance of objectives in technical development<br />

Sequence Objective Activity<br />

1. Maximize product utility. Increase efficiency and effectiveness of the product for the<br />

customer.<br />

2. Minimize raw material costs. Reduce material consumption, waste, and expensive substances.<br />

3. Minimize labour costs. Reduce staff, reduce wages, and optimize external consulting<br />

effort.<br />

4. Minimize tax payments. Optimize distribution of costs, use government support measures.<br />

5. Optimize maintenance intervals.<br />

Adapt product to service infrastructure and competences.<br />

6. Minimize supply risks. Optimize size of supply network and tier co-operation.<br />

7. Optimize capital management. Adapt production to own capital, shareholder interest etc.<br />

8. Minimize novelty conflicts. Reduce efforts for product implementation and version changes.<br />

9. Optimize ecological impact. Reduce energy consumption, hazardous waste etc.<br />

10. Optimize intellectual property. Reduce risk of copies by competitors, dependency on outside<br />

patents, knowledge etc.<br />

Assuming such a scenario of research and development activities, technical change and the decision about<br />

its profitability must be understood as two separate procedures. If so, it makes sense to approach these<br />

activities in terms of evolution. The creation of solutions and their evaluation and selection for further use<br />

can be considered to interact as displayed by the boxes in Figure 1. It is important to note that the specific<br />

expertise in marketing and accounting and in research and development does not contradict the idea of<br />

evolutionary adaptation. In fact, the high level of sophistication of the work in the different departments<br />

must be seen as the very reason why it is necessary to use an evolutionary approach: it is impossible to<br />

combine all the expertise in both fields. The complexity of the task is so high that it requires a separation.<br />

The success of the technical development as an overall optimization depends on various parameters. In<br />

particular, the way how the change operators are applied has to be appropriate to the problem. Market<br />

demand in general is not determinate enough to reflect this issue. The dimensions of the investment calculations<br />

in accounting, however, provide very explicit information about the evaluation of the solution<br />

quality. In that sense, the role of management accounting for the selection of technical solutions is easily<br />

comparable to the role of a formal evaluation function in a computational multi-dimensional optimization<br />

problem. Recent research on the design of Genetic Algorithms can therefore convey important insight into<br />

the question of innovation management.<br />

4. Evolutionary models and engineering expertise<br />

4.1 Operators and landscapes<br />

Goldberg discusses the task of innovation management as the design of a competent Genetic Algorithm. In<br />

this context, competence means that the algorithm is able to identify good solutions. For Goldberg, this<br />

depends on the question whether the operators are able to reproduce the good characteristics of previous<br />

alternatives (Goldberg 1998). Topologically, the problem can be described as the question whether it is<br />

easy to find a path from any given alternative to the best ones. This problem is addressed by the notion of<br />

fitness landscapes.<br />

A fitness landscape is a triple (S, q, N) of a search space S, an evaluation function q and a neighbourhood<br />

structure N on that space. The neighbourhood structure is induced by the changes that an operator can<br />

cause for a given element of the search space, for example by all the mutations that can be performed on<br />

an alternative. For certain instances of the Travelling Salesman Problem, (Boese et al. 1994) have shown<br />

that the landscape induced by a mutation operator on an intuitively plausible metric leads to a structure<br />

shaped like a big valley on the search space. With respect to that landscape, local optima are situated close<br />

200


Albrecht Fritzsche<br />

to one another and a simple heuristic search algorithm can calculate best solutions quite effectively.<br />

(Reeves 1999) has extended this approach to a permutation flow shop scheduling problem using operators<br />

for mutations, shifts and inversions of parts of the sequence. With reference to these operators, the landscape<br />

on the search space provides easy paths to local global optima. (Knowles and Corne 2002) discuss<br />

different strategies of search on a global assignment problem, developing solutions intently either in the<br />

direction of better quality or somewhat "horizontally" to that direction once that good solutions have been<br />

found.<br />

In a complex multi-objective optimization problem, the direction of change that leads to better quality is<br />

generally unknown. However, the single objectives suggest certain directions in which intentional change<br />

can proceed. With respect to one dimension of the problem, these directions indicate the steepest ascent of<br />

solution quality. Under the assumption that the different objectives of the problem are not fully contradictory<br />

in their notion of good solutions, changes in these directions can be expected to have strong influence on<br />

the overall fitness, too. It therefore seems sensible to introduce operators that work explicitly on the improvement<br />

of solutions with respect to single objectives.<br />

Recent findings about the application of heuristic search methods to order scheduling problems indicate<br />

that such operators can indeed have a positive effect on the optimization procedure (Fritzsche 2009, see<br />

also Solnon et al 2008). Genetic Algorithms for order scheduling show better results, if they do not only<br />

execute arbitrary mutations and re-combinations but also constructive changes of the schedule with respect<br />

to the given objectives. For example, it is useful to apply operators which can move items in the<br />

sequences according to given constraints of minimal production distances or the due dates of the orders.<br />

With reference to such constructive operators, the notion of expertise can be introduced to the model of<br />

evolutionary adaptation.<br />

4.2 Expertise in evolutionary procedures<br />

Based on the study of fitness landscapes, the role of expertise in technical development can be explained<br />

in terms of the directions in which intentional changes are performed. Management accounting defines<br />

these directions by setting up the dimensions of the profitability model. The dimensions give orientation for<br />

the efforts to improve technical solutions. Table 2 gives an overview over the different activities that contribute<br />

to technical change with respect to the dimensions of profitability. These activities reflect the diversity<br />

of the work in research and development as well as some other departments that are closely connected<br />

to it. While Goldberg's model of technical change only involves two different operators, it seems more<br />

adequate to consider the contributions of product designers, project managers, physicists, technicians and<br />

engineers as many separate operators. All of them contribute in different ways and to a different extent to<br />

improvement with respect to the single objectives.<br />

Table 2: Expertise involved in technical development<br />

No. Objective Activity<br />

1. Maximize product utility. Product design and usability, marketing, social studies.<br />

2. Minimize raw material costs. Physics, material research.<br />

3. Minimize labour costs. Factory design, robotics, training departments.<br />

4. Minimize tax payments. Project management, infrastructure planning.<br />

5. Optimize maintenance intervals. Customer relationship management and service.<br />

6. Minimize supply risks. Modular design, cross-company research.<br />

7. Optimize capital management. Financial advice.<br />

8. Minimize novelty conflicts. Change management and engineering.<br />

9. Optimize ecological impact. Ecological auditing, systems engineering.<br />

10. Optimize intellectual property. Legal departments, R&D administration.<br />

The comparison of these activities with the constructive operators in Genetic Algorithms leads to an important<br />

conclusion about the way how these activities facilitate technical improvement. They do not have to<br />

be brought together in a systematic process of rational problem solving; technical improvement is also<br />

possible in a more liberal scenario of interaction. If all different experts are free to contribute to the solutions<br />

in whichever way they like, they can still evolve into something better, even if the single changes performed<br />

201


Albrecht Fritzsche<br />

on the solutions seem highly contradictory. Innovation management, however, has to make sure that this<br />

process takes place under a certain selective pressure of profitability. In addition, it is necessary to make<br />

sure that the changes of the solutions do not become too big, because this would inhibit the interaction<br />

between the operators. Small contributions to change, like the measures for continuous improvement in a<br />

factory, can therefore play a huge role for evolutionary adaptation.<br />

4.3 Expertise and the market<br />

Genetic Algorithms using constructive operators show how technical development can proceed as an<br />

evolutionary process in the presence of professional expertise. A fundamental problem of the whole approach,<br />

however, has so far remained unmentioned: If technical development is subject to the pressure<br />

exerted by management accounting, the solutions it creates can only be as good as the model of profitability<br />

that is used in the investment calculations. Since this model can never be fully adequate to the reality<br />

of market demand, it must be assumed that technical change always leads into a slightly wrong direction. It<br />

will not adapt to market itself, but to the dimensions in which the profitability is measured. Technical solutions<br />

will consequentially reflect all the strengths and weaknesses of management accounting. In particular,<br />

the development will be detrimental with respect to all aspects of market demand that are neglected<br />

in the investment calculations. If so, the development of technology can never fully meet the economic<br />

necessities. In fact, one would have to expect that it is often led far astray.<br />

The study of fitness landscapes, however, indicates that the evaluation function is not the only source of<br />

influence on the direction of an evolutionary process. The neighbourhood topology also plays an important<br />

role. The search algorithm can only identify solutions that are within reach of the operators applied under<br />

the circumstances set by the population size and the selection technique. In that sense, the operators do<br />

not only change the course of the search; they also set the boundaries of the areas in which the search<br />

takes place. If the search space is very big, it is quite important to reduce the search procedure to a suitable<br />

subset of the space in order to identify good solutions. If the operators perform very specific changes, the<br />

subset in which the search proceeds can be rather small. The application of such operators is justified, if<br />

they can be assumed to cover a subset of the search space that contains the elements with the highest<br />

quality. Operators that perform changes in the direction of the dimensions of the evaluation function can be<br />

expected to do so, if they are used interactively.<br />

Considering this, it becomes clear that the direction of evolutionary processes is not determined exclusively<br />

by the selective pressure under which they take place. The selection of the best elements can only cause<br />

an adaptation in the boundaries set by the operators that generate new alternatives. It therefore would be<br />

wrong to say that technical development is completely dominated by management accounting. Before it is<br />

possible to select the best technical solutions, they first have to be generated by research and development.<br />

Although the activities involved in research and development are related to certain dimensions of the<br />

profitability model in accounting, there is good reason to assume that they are also influenced by the<br />

common sense of the people involved. This influence becomes visible, for example, in the variations of<br />

technical development between different cultures. Although one can assume that the profitability calculations<br />

in competing companies are rather similar, the technical solutions that are generated in the various<br />

<strong>European</strong> countries, America or China are quite diverse. It lies at hand to attribute these variations to the<br />

fact that work in research and development is not solely guided by professional expertise in certain fields,<br />

but also by general assumptions about profitable technical solutions.<br />

5. Conclusion<br />

The purpose of this paper has been to show how different activities in a company can be brought together<br />

in an evolutionary model of technical change. Research on Genetic Algorithms gives evidence that systematic<br />

procedures to define and execute improvement do not contradict the evolutionary approach. Industrial<br />

implementations of Genetic Algorithms often use sophisticated evaluation functions that bring together<br />

many different objectives. In many cases, specific operators are used to identify better solutions with<br />

respect to one single objective. These operators can significantly increase the performance of the algorithm,<br />

if the interaction between them is organized in the right way.<br />

David Goldberg has repetitively claimed that the way how Genetic Algorithms produce change is closely<br />

connected to human skills in technical development (e.g. Goldberg 1998, Goldberg 2002). This paper does<br />

not give evidence that supports this claim, but it makes clear that the evolutionary search in Genetic Algorithms<br />

can serve as a reference for a model of technical development in a complex economic context.<br />

The concept of Genetic Algorithms gives an explanation how the investment calculations in management<br />

202


Albrecht Fritzsche<br />

accounting and the technical change procedures in research and development can interact as two separate<br />

processes. In addition, Genetic Algorithms using sophisticated operators show how it is possible that<br />

the direction of technical improvement deviates from the selective pressure of profitability. In the future,<br />

studying Genetic Algorithms may allow further conclusions about innovation management.<br />

References<br />

Beck-Sickinger, A.G. and Petzold, M. (ed.) (2009) Paradigma Evolution, Frankfurt, Peter Lang.<br />

Boese et al. (1994) "A new adaptive multi-start technique for combinatorial global optimizations", Operations Research<br />

Letters 16, Amsterdam, Elsevier, pp 101-113.<br />

Broesel, G., Keuper, F. and Woelbing, I. (2007) "Zur Uebertragung biologischer Konzepte in die Betriebswirtschaft",<br />

Zeitschrift fuer Management, 2(4), pp 436-466.<br />

Dennett, D.C. (1996) Darwin’s Dangerous Idea: evolution and the meanings of life, New York, Simon & Schuster.<br />

Dopfer, K. (2008) "Was ist Evolutionsökonomie?", Innovation zwischen Markt und Staat (Ebner, A., Heine, K. and<br />

Schnellenbach, J. ed.), Baden-Baden, Nomos.<br />

Frenken, K. (2006) Innovation, Evolution and Complexity Theory. Cheltenham/ Northampton, Elgar.<br />

Fritzsche, A. (2009) Heuristische Suche in komplexen Strukturen, Wiesbaden, Gabler.<br />

Goldberg, D.E. (2002) The Design of innovation: Lessons from and for competent genetic algorithms, Boston, Kluwer<br />

<strong>Academic</strong>.<br />

Goldberg, D.E. (1998) The Race, the Hurdle, and the Sweet Spot, (IlliGAL Report 98007), Chicago, Univ. of Illinois.<br />

Hodgson, G. M. (2005) “Decomposition and Growth: Biological Metaphors in Economics from the 1880s to the 1980s”,<br />

The Evolutionary Foundation of Economics (Dopfer, K ed.), Cambridge, Cambridge University Press,<br />

pp.105-150.<br />

Hodgson, G. M. (1993) Economics and Evolution: Bringing Life Back Into Economics, Cambridge, Polity Press.<br />

Klaus, P. (2005) "Die Frage nach der optimalen Komplexität in Supply-Chains und Supply-Netzwerken", Perspektiven<br />

des Supply Managements, (Essig, M., ed.). Berlin/ Heidelberg, Springer, pp363-376.<br />

Knowles, J. and Corne, D. (2002) "Towards Landscape Analyses to Inform the Design of Hybrid Local Search for the<br />

Multiobjective Quadratic Assignment Problem", Soft Computing Systems: Design Management and Application,<br />

Amsterdam, pp 271-270.<br />

Marshall, A. (1890) Principles of Economics, Reprint 1930, London, Macmillan.<br />

Nelson, R. R. and S. G. Winter (1982) An evolutionary theory of economic change, Cambridge, Belknap.<br />

Reed; P.M., Minsker, B. S. and Goldberg, D. E. (2001) "The Practitioner’s Role in Competent Search and Optimization<br />

Using Genetic Algorithms", Bridging the Gap. (Phelps, D. and Shelke, G. ed.), <strong>Conference</strong> Proceedings, Reston,<br />

ASCE, 1, 341.<br />

Reeves, C. R. (1999) "Landscapes, operators and heuristic search", Annals of Operations Research 86, 1999, pp<br />

473-490.<br />

Schumpeter, J. A. (1912) Theorie der wirtschaftlichen Entwicklung, Berlin 1911; Reprint 2006, Berlin: Duncker &<br />

Humblot.<br />

Simon, H. A. (2005) “Darwinism, Altruism and Economics”, The Evolutionary Foundation of Economics (Dopfer, K ed.),<br />

Cambridge, Cambridge University Press, pp 89-104.<br />

Solnon, Ch, Cung, V.D., Nguyen, A. and Artigues, Ch (2008) "The Car Sequencing Problem", <strong>European</strong> Journal of<br />

Operational Research 191(3), pp 912-927.<br />

Trott, P. and Lataste, A (2005) “The Role of Consume Market Research in New Product Decision Making: Some Preliminary<br />

Findings”, The Entrepreneurship-Innovation-Marketing Interface (Gaul and Jung ed.), Künzelsau, Würth.<br />

Veblen, T. B. (1898) "Why Is Economics Not an Evolutionary Science?", Quarterly Journal of Economics, 12(3), July,<br />

pp 373–97.<br />

Winter, S. G. (2005) “Towards an Evolutionary theory of Production”, The Evolutionary Foundation of Economics<br />

(Dopfer, K ed.), Cambridge, Cambridge University Press, pp 223-254.<br />

Wolpert, D. H. and Macready, W. G. (1997) "No Free Lunch Theorems for Optimization", IEEE Transactions on Evolutionary<br />

Computation 1.<br />

203


Case Study on Information Evaluation by GIS for Aging<br />

Society Urban Planning: GIS Application on Urban Planning-<br />

Hiroatsu Fukuda, Yupeng Wang and Kiyoshi Shinriki<br />

The University of Kitakyushu of Japan, Japan<br />

fukuda@env.kitakyu-u.ac.jp<br />

y09e0101@hibikino.ne.jp<br />

shinriki@hq.kiu.ac.jp<br />

Abstract: Japan’s aging society and shrinking population are leading to serious problems, including a declining level<br />

of community vitality in most of Japan’s secondary cities. Structural changes in Japan’s population are expected to<br />

continue for the near future, making sustainable urban design and renovation more important than ever. Proper<br />

urban design must seek and evaluate the relevant social information. This research presents a typical case study<br />

using a new information evaluation method, the Geographic Information System (GIS), against the background of an<br />

aging and shrinking society. It is carried out in Yahatahigasi-ku, a ward in the city of Kitakyusyu, Japan. The ward’s<br />

land is characterized by steep slopes, which are inconvenient for aged people and reduce their outdoor mobility.<br />

Moreover, most of the ward’s houses are over 30 years old, the average lifespan for houses in Japan. The quality of<br />

the houses is inadequate to the demands of modern life, perhaps explaining the area’s inability to retain its youth. As<br />

a result, the shopping districts of Yahatahigasi-ku, prosperous in the 1950s and 1960s, are declining in proportion to<br />

the number of deserted houses This research explores GIS design approaches as configured for an aging society. It<br />

discusses the important factors that flow into an urban design meant to address a declining population, such as the<br />

plot ratio, the mixing of commercial and residential spaces, and the placement of public facilities for the aged. The<br />

city planner must find the most suitable area for planning. Sustainable urban design requires that new elements be<br />

introduced into the traditional configuration; the constructional style of buildings in a mixed-use area will reflect the<br />

character of the district. This study considers the GIS urban planning method as a development of the GIS<br />

application. We see here the importance of geographic information to urban planning. This paper will exemplify the<br />

importance of sharing informational evaluations to the urban planning process. In this paper, a case study of Japan<br />

was investigated. It makes contribution on urban regeneration for aging society that also appropriate for cities outside<br />

Japan.<br />

Keywords: urban design, GIS, aging society, compact city, migration<br />

1. Introduction<br />

Urban design tends to correspond to the development of the city. One consequence of Japan’s declining<br />

population is that its shrinking housing density is reducing the utilization of cities’ public facilities. The<br />

results of GIS condition analyses allow us to identify the sites most in need of innovation. Information<br />

evaluation by GIS would be a crucial component of this planning method. This study proposes that<br />

reducing the size and compacting the functions of its cities would solve Japan’s social problems. Efforts<br />

to cope with reduced utilization efficiency by reducing city sizes have not been well executed but<br />

represent an important facet of dealing with our globally aging society. This proposal should be qualified<br />

to meet the special requirements of the aged, who tend to retain a traditional attitude towards the street<br />

that may no longer be suitable. Other, newer elements are also indispensable to sustainable city<br />

planning. These new elements could evolve from traditional ones, connecting different generations. This<br />

study analyzes factors such as the plot ratio, the building type, the mixing of commercial and residential<br />

spaces, and the locations of public facilities for the elderly.<br />

2. Background and goal<br />

Aging society and the diminishing population are getting more serious in Japan, leading to related<br />

problems, such as the declining vitality of communities in most of the secondary cities. This tendency of<br />

the population and its structural change is supposed to continue in near future, thus makes the<br />

sustainable urban design urgent.<br />

Japanese average life span is 78.6(men), 85.5(women) in 2007 the longest in the world. The population<br />

ratio over 65 years old is 22.8% in 2009, largest in the world. As the prospects in 2055, by the research<br />

conducted by Japanese government in 2005, the population will decreased from 127.8 million to 89.9<br />

million. The average life span will be 83.7(men), 90.3(women). The population ratio over 65 years old will<br />

be 40%. According to this result, aging society problem will grow into a serious problem in the further.<br />

The industrial structure of Yahatahigasi-ku has totally changed over the last 50 years in tandem with<br />

changing lifestyles. In 1960 to 70, Kitakyushu supported Japanese rapid economical growth, through the<br />

204


Hiroatsu Fukuda et al.<br />

aggregation of such material industries as the steel, chemical, metal and ceramic industries, which began<br />

when the government-managed Yahata Steelworks, the largest steelworks in Asia, which was founded in<br />

1901. Yahatahigasi-ku, the biggest industrial area of Japan, was deeply affected by the changing times.<br />

Today, the material industry moved to the Asian countries, high-technology industries including industrial<br />

robots, IC-related products and biotechnologies are being generated. The shopping streets of<br />

Yahatahigasi-ku, prosperous in the 1960s and 1970s, are in decline, as the proportion of deserted<br />

houses increases; a new plan for urban regeneration is expected. Yahatahigasi-ku’s problem exists<br />

throughout Japan and other developed countries. An approach to the reconstruction of these typical old<br />

towns will be discussed in this paper.<br />

3. Methods of research<br />

We will first analyze the existing condition through GIS and propose a construction site based on a<br />

comprehensive understanding of the site information evaluation and contracting studies.<br />

The date of population was obtained from web side of Kitakyushu city. The landform current situation and<br />

the distribution of living facilities of Yahatahigasi-ku were inspected by field trips. All of the dates are<br />

collected and synthesized by GIS<br />

The building typologies with new elements appropriate for the elderly will be proposed.<br />

4. Case study<br />

Yahatahigasi-ku is a ward in the city of Kitakyushu, where Yahata Steelworks is located and the people<br />

who work for the company live. The population here is decreased from 81,956 (1988) to 73,162 (2008)<br />

and the population ratio over 65 years old is 30.2% now. The land here is characterized with the steep<br />

slop which is inconvenient for aged people and reduces the chances of their outdoor activities.<br />

On the other hand, most of the houses here are used far more than 30 years, the average lifespan for<br />

houses in Japan. The quality of the house is not adequate for the modern life which may be responsible<br />

for the losing of young people.<br />

4.1 City size reduction<br />

Figure 1: Overview of Yahatahigasi-ku<br />

205


Hiroatsu Fukuda et al.<br />

As shown in the overview of Yahatahigasi-ku, most areas of this ward are mountainous, with steep<br />

slopes. Many residential areas were built for employees during the period of Japan’s industrialization.<br />

Therefore, housing developments extended to the steeply sloped areas as the city grew. The proportion<br />

of retirees has increased over the last 50 years; meanwhile, young people have been fleeing the city<br />

because of its unsuitable living conditions. Thus, the concept of “compact” is an appropriate response to<br />

the ward’s decline. First, we must designate a proper site.<br />

As shown in Figure 2, the steeply sloped areas on the mountain pose an acute problem: access by car is<br />

impossible; one can only walk after the slope begins, making it hard for inhabitants to perform daily<br />

outdoor activities. The elderly, in particular, who have to get to nursing facilities and go shopping, cannot<br />

bear the necessary effort.<br />

Figure 2: Pictures of the residences on the steeply sloped area<br />

This ward has five cultural zones. Two of those are on the steeply sloped area, as shown in Figure 2, and<br />

three lie on a flat area. Four main roads cross in the central zone; there is a railway station near the flat<br />

area that is conveniently accessed from outside the ward. There is a famous theme park, “Space World,”<br />

the biggest commercial area in this ward, and an area boasting three museums and other public facilities.<br />

City functions are already densely compacted here. Based on its merits, the central zone on the flat area<br />

was selected as the best site for innovation.<br />

4.2 Inhabitant migration proposal with information evaluation by GIS<br />

Figure 3 shows the designated migration-in promotion area, migration-out promotion area, and the buffer<br />

area. The study proposes an inhabitant migration from the sloped area to the flat area based on an<br />

examination of the aging population and the old dwellings on the slope as analyzed by the GIS. Many<br />

residences were built on the steeply sloped area. To solve the environmental problems produced by the<br />

steep slope, flat areas were chosen as the object region for inhabitant migration. The migration is<br />

expected to reduce the size of city and compact it, improving the efficiency of city services and facilities.<br />

Migration-out promotion areas are those with high slopes and high ratios of elderly—where there is,<br />

roughly, a level higher than 80m and a slope degree higher than 5. We propose to keep inhabitants from<br />

moving in and to encourage them to move out of this area.<br />

Migration-in promotion areas are flat with convenient traffic and public facilities—where there is, roughly,<br />

a level lower than 40m and a slope degree lower than 3. In addition the areas where is anticipated to rich<br />

of these facilities in the further.<br />

Buffer areas lie between migration-out promotion areas and migration-in promotion areas. Government<br />

offices and universities are situated there. Migration-out areas will become sparsely populated during the<br />

process of migrating-in. In the buffer areas, sparse landscape, community decay, and crime in streets full<br />

of abandoned houses could be alleviated by adding public facilities and encouraging more human<br />

interaction.<br />

A high density of population and buildings is required to keep public facilities efficient. Inhabitants living in<br />

migration-out promotion areas cannot access public facilities conveniently; meanwhile, those public<br />

206


Hiroatsu Fukuda et al.<br />

facilities will decay as population density decreases. Sustainable city planning must compact city<br />

functions and increase population density.<br />

Figure 3: Object region for inhabitant migration<br />

Figure 4: The situation of public facilities<br />

207


4.3 Share-ride taxi<br />

Hiroatsu Fukuda et al.<br />

The share-ride taxi is a kind of conveyance that can take a maximum of 10 people a short distance. It<br />

normally services only one or two residential areas. It does not use a designated station and can stop<br />

and collect passengers anywhere along its route. The price of a share-ride taxi is lower than that of a<br />

normal taxi and is more convenient for the aged than the bus. It can be used for daily shopping or other<br />

short errands. We can expect this form of sharing to lighten the city’s traffic. This system has been<br />

successfully introduced into many areas of Japan.<br />

Figure 5: Pictures of share-ride taxi (http://www.asagiri-town.net/q/aview/102/687.html)<br />

Two share-ride taxi routes are proposed for the central zone. Route 1 is a circle connecting a big hospital<br />

and the old residential areas with the steep slope on the mountain. Route 2 is a circle connecting a big<br />

shopping centre, three museums, and the theme park. Route 1 and 2 cross at the mixed-use area. The<br />

operational hours are to be from 9 am to 5 pm. The full route will take around 20 minutes to travel. It is<br />

now used mostly by housewives and students.<br />

Most of the facilities on the routes could not be reached by bus from the mixed-use area. Life would be<br />

made more convenient for the aged if their public facilities were connected. The GPS (Global Position<br />

System) allows users to pinpoint the location of their taxi in real time on their mobile phones.<br />

Figure 6: Routes of share-ride taxi<br />

208


Hiroatsu Fukuda et al.<br />

Figures 7 to 9 show the image of each stage of the Share-Ride Taxi Rolls. Initially, share-ride taxis will<br />

connect the mountainous areas with the flat commercial areas.<br />

Figure 7: The first stage of share-ride taxi rolls<br />

Figure 8: The second stage of share-ride taxi rolls<br />

Figure 9: The third stage of share-ride taxi rolls<br />

209


Hiroatsu Fukuda et al.<br />

After 10 years, the population density of the flat area will increase with the addition of the inhabitants from<br />

the sloped area. The taxi route in the flat area is proposed for the elderly and others (like students) who<br />

cannot drive.<br />

After 20 or 30 years, the taxis will connect several central regions in the flat areas. Given the results of<br />

the inhabitant migration program, routes connecting the mountainous and flat areas will not be<br />

necessary. Taxi routes in the flat area should enhance the convenience of the compacted city area.<br />

4.4 Site analysis<br />

Figure 10: Public facilities Figure 11: The zoning of city functions<br />

The central zone used to be a gateway into the residential area, through which employees would pass on<br />

their way home from work. We will explore the character of the central zone, analyze the relative<br />

positions of some of its facilities, and discuss the city’s zoning function.<br />

The central zone mixes commercial and residential uses. In this mixed-use area, all necessary city<br />

facilities are located within 500m of the center and are thus conveniently accessible. This mixed-use area<br />

is therefore suitable for habitation.<br />

A city axis extends from the station, goes through the mountainous area, and passes through the<br />

mixed-use area. There should also be a green axis extending from the mountainous area, in order to<br />

strike a balance between the city and nature.<br />

As an urban planning restriction, the maximum plot ratio for the mixed-use area is 500%. However, the<br />

actual plot ratio is much lower than that, as it does not take full advantage of existing city functions. To<br />

maintain a bustling city community, both the population density and plot ratio should be increased.<br />

The mountain at the south side of the site is key to the landscape. We propose to take advantage of the<br />

view of that mountain. This can be achieved through the appropriate building disposition and<br />

building-type analysis.<br />

210


Figure 12: Restriction of plot ratio<br />

4.5 Building-type analysis<br />

Hiroatsu Fukuda et al.<br />

The maximum plot ratio for the mixed-use area is 500%. To achieve a residential building style through<br />

sustainable urban design, we must consider three functions of the building under the plot ratio restriction.<br />

Figure 13: Land use zoning in 3D space<br />

211


Hiroatsu Fukuda et al.<br />

Commerce: commerce affects the activity on the street and should thus occur on the lower floors, not<br />

only for the sake of daily shopping but also for amusement. The main target is youth who live elsewhere.<br />

The plot ratio for commerce is about 100%.<br />

Service facilities for residences: considering our aging society, service facilities for the local elderly are<br />

located on the higher floor: this makes them easier to access and somewhat separate from activities on<br />

the street. The plot ratio for residential service facilities is about 100%<br />

Residences: we will draw contrasts among 5 building types in order to discuss the way the residential<br />

environment is linked to the plot ratio, the number of plies, and the construction styles. We expect to see<br />

high buildings with low plot ratios.<br />

Three types of relationship between building and street are contrasted.<br />

Figure 14: Building types<br />

Effective ventilation, solar irradiation, and privacy can be ensured by changing the residence into a slim<br />

tower; this way, the public space will be used only by residents.<br />

Type 1: face onto the street<br />

Commerce on the lower floors is close to the street. The facade of the Japanese traditional street can be<br />

maintained.<br />

Type 2: Increase greenery<br />

Set the lower floors back and introduce a green area in an open space, though this would differ from the<br />

traditional Japanese style. A slim building also increases the amount of sky an inhabitant can see from<br />

the street. Type 2 takes advantage of the city’s rich and varied local character.<br />

5. Conclusion<br />

This paper analyzed a method of urban design for an aging society through a case study that applies GIS<br />

to a Japanese city, Yahatahigasi-ku, once considered a leading city during Japan’s industrial age. A GIS<br />

information evaluation leads us to propose a method of urban reduction that requires compacting city<br />

functions, connecting access points to public facilities, and improving the residential environment.<br />

Introducing GIS is crucial to achieving the goal of designating a suitable site for inhabitant migration. The<br />

ideal site is the central zone, once an important ward gateway.<br />

212


Hiroatsu Fukuda et al.<br />

Therefore, it is imperative that the city planner find the most suitable area for planning by way of a site<br />

information evaluation. Additionally, formulating a sustainable urban design requires that new elements<br />

be introduced into the traditional configuration. Ultimately, the constructional style of buildings in<br />

mixed-use area will reflect the character of their district.<br />

Acknowledgments<br />

The authors wish to express their appreciation to “Revitalization of Local Area Project” for providing the<br />

subvention for this research.<br />

References<br />

Dewanchker B., (1997) “Comparison of the Change in Industrial Development in the Kitakyushu Industrial Zone and<br />

the EMSCHER Zone in Germany”, Journal of architecture, planning and environmental engineering.<br />

Transactions of AIJ (502) pp.51-56<br />

Dewanchker B., Takahasi N. and Ojima T., (1997) “Ecological Improvement and the Redevelopment of an Abandon<br />

Industrial Site in the City of Uozu”, Summaries of technical papers of Annual Meeting Architectural Institute of<br />

Japan. pp.821-822<br />

Fukahori H. (2006), “Disaster prevention map making with IT method”, Journal of Japan Sewage Works Association,<br />

43(527) pp.43-46<br />

Fukahori H. (2007), “Make Kitakyusyu-city an nature environment capital”, Journal of sewerage, monthly 30(2) (422)<br />

pp.38〜41<br />

Gao W., Ariyama T. and Ojima T., (1999) “Prediction on the Energy Consumption of Renewing the Residential<br />

Recycle Material”, Journal of architecture, planning and environmental engineering. Transactions of AIJ (516)<br />

pp.101-106<br />

Gao W. and Ojima T., (1995) “Study on Planning an Ecological City in SHITAMACHL Tokyo”, Journal of architecture,<br />

planning and environmental engineering. Transactions of AIJ (472) pp.63-71<br />

Gao W., (1993) “Thermal Effects of Open Space with a Green Area on Urban Environment: Part I: A theoretical<br />

analysis and its application”, Journal of architecture, planning and environmental engineering. Transactions of<br />

AIJ (448) pp.15-27<br />

Song H., Guan D., Fukahori H. and Gao W., (2009) “Study on utilization of semi-industrial districts based on GIS in<br />

Kitakyushu : Investigation of present status of semi-industrial districts in Kitakyushu Part 1”, Journal of Asian<br />

Architecture and Building Engineering, pp.529-532<br />

Urban planning general overview of Kitakyusyu city of Japan, (2007) Planning department in city urban planning<br />

bureau of Kitakyusyu, Japan<br />

213


Distortion Free Algorithm to Handle Secondary Watermark<br />

Attack in Relational Databases<br />

Sajid Iqbal, Azhar Rauf, Huma Javed and Shabir Ahmad<br />

University of Peshawar, Pakistan<br />

sajidiqbal84@gmail.com<br />

rauf.azhar@yahoo.com<br />

humajaved15@yahoo.com<br />

shabir_usmany@yahoo.com<br />

Abstract: Publishing data over the internet has greatly facilitated the ways of data transfer and file sharing but<br />

opened new threats of piracy and copyright violations. The technique of relational database watermarking is used to<br />

protect digital assets by embedding imperceptible watermarks into the data before publishing. Relational database<br />

watermarking got inspiration from digital watermarking and efforts are being made to protect databases from such<br />

attacks by embedding watermarks. Database watermarking has been an active area of research for the past few<br />

years. This process causes distortion of data and may affect the quality of information and reporting. A novel<br />

approach is proposed to embed non-visible Unicode character spaces as watermark instead of changing the least<br />

significant bits of numeric data, which avoids distortion of the data. Similarly in the previous works, watermarks were<br />

embedded in numerical data only while our technique is applicable to character type data which is another novelty of<br />

the proposed technique. Researchers have identified different types of attacks against the databases. One of them is<br />

secondary watermark addition attack. The proposed technique provides a distortion free solution to protect the<br />

database from secondary watermark addition attacks. The owner’s private key along with a timestamp issued by a<br />

trusted timestamp authority is embedded to protect the database from the threats of secondary watermark addition<br />

attack. Instead of involving the third party at a later stage, we propose its involvement at the stage of watermark<br />

construction. The third party will keep the record of timestamps, issued against owner’s encrypted key, and database<br />

for which the timestamp has been issued. Experiments have shown that our algorithm is effective even after changing<br />

90% of the contents of database by the attacker. The proposed technique not only takes care of the secondary<br />

watermark addition attack but also other types of attacks for example subset attacks and bit attacks.<br />

Keywords: relational database watermarking, secondary watermark, trusted time stamping<br />

1. Introduction<br />

In the current age of Technology, data is on the move. People can transmit data over the internet or copy<br />

it from different locations. The protection of digital assets like images and videos from piracy and<br />

copyright protection is done through watermarking. But to protect databases from piracy, little effort has<br />

been done so far.<br />

Watermarking Relational databases got its inspiration from digital watermarking. In relational databases,<br />

watermarks are embedded in the least significant bits of numeric values of some specific columns of<br />

selected records (Agarwal 2002) (Hu 2009) (Zhang 2004) (Li 2005) (wang 2008) (Cui 2008). The<br />

selection of these records and columns is based upon some hash function applied to owner’s defined<br />

key, image, voice or fingerprint (Agarwal 2002) (Hu 2009) (Zhang 2004) (Li 2005). Information hiding<br />

using watermarking technique causes some distortion but it is assumed that this distortion is acceptable<br />

(Agarwal 2002) (Hu 2009) (Zhang 2004) (Li 2005) (Wang 2008) (Cui 2008). Later on, if at some stage the<br />

real owner suspects the piracy of his database, he can use the watermark detection algorithm to recover<br />

the hidden pattern and thus prove his ownership.<br />

In this paper we present a better approach to watermark relational databases. We thus propose a novel<br />

distortion free algorithm to protect relational database assets from reversibility and secondary watermark<br />

attacks. In our proposed scheme, an identification image is used as a watermark embedding information<br />

which proves the copyright information. The proposed algorithm is resistive to different kinds of attacks<br />

including benign updates, malicious attacks, reversibility attack and secondary watermark attack.<br />

Rest of the paper is organized as: Section 2 gives a brief introduction to secondary watermark addition<br />

attack. Section 3 describes the existing work regarding secondary watermark attack in relational<br />

database watermarking. Section 4 describes the notations used and the proposed algorithm for<br />

watermark insertion and detection. Section 5 shows some experimental results and analysis and finally<br />

section 6 concludes this paper.<br />

214


2. Secondary watermark addition attack<br />

Sajid Iqbal et al.<br />

Watermarking databases and digital assets has gained enormous attention from researchers since the<br />

term was coined by Agarwal et al. (Agarwal 2002). One of the many reasons behind this effort is the fast<br />

movement of data across internet and many file sharing services for sharing and transfer.<br />

So far, different types of attacks have been observed against the relational databases causing piracy and<br />

copyright conflict including benign updates and malicious attacks. One of the important attacks in this<br />

regard that causes ownership conflict is secondary watermark addition attack. In secondary watermark<br />

attack, the attacker simply successfully discovers an unreal watermark and then fraudulently embeds his<br />

own watermark. Using that watermark he can claim his ownership, thus causing ownership conflict. It<br />

cannot be confirmed which watermark was embedded first. The scenario of secondary watermark attack<br />

can be shown as follows:<br />

The real owner of relation embeds his watermark WR to the original Relation RO using some private key K<br />

as:<br />

WE (WR, K, RO) Watermarked Relation where WE is a function that embeds watermark to the<br />

relation.<br />

An attacker finds an unreal watermark WR' although he doesn’t have access to the watermarked relation.<br />

He embeds a false secondary watermark to the database in order to prove his ownership. The addition of<br />

secondary false watermark to an already marked database can be shown by a scenario as follow:<br />

WE' (WR', K', Rw) Watermarked Relation<br />

Where WE' is the fraudulent operation that embed unreal watermark to the marked relation Rw.<br />

Now the relation contains the watermark of two different persons, the real owner and the attacker. This is<br />

hard to prove the ownership of a relation as both the persons can run their algorithm and thus detect their<br />

respective watermark.<br />

Embedding a secondary watermark to a watermarked relation can cause Destruction.<br />

3. Related work<br />

Relational database watermarking for copyright protection and piracy protection is one of the major<br />

research areas now a day. Different approaches have been adapted to tackle with piracy attacks. But so<br />

far there has been no prominent achievement to handle the secondary watermark addition attack in<br />

relational database.<br />

(Dong 2009) proposed the insertion of characteristic code as watermark information to take care of<br />

invertibility attack. In their proposed approach, the characteristic code is calculated by selecting an even<br />

number of attributes from the relation and calculating the average value of each attribute. Next the even<br />

number of attributes is divided in pairs and members of each pair are compared with each other. And<br />

finally string of bits to be watermarked is calculated by taking XOR of compared attributes. When<br />

ownership conflict occurs, owner can prove his ownership by the help of characteristic code. But a very<br />

obvious shortcoming with this approach is that in order to prove the ownership one must have original<br />

database relation which goes against one of the primary characteristics of database watermarking<br />

“Blindness” as defined by (Agarwal 2002)<br />

(Gupta 2009) proposed another watermark technique, resilient to secondary watermark addition attack.<br />

They claim that their technique is resilient to benign updates, additive attacks and bit attacks. They insert<br />

watermark information to float value attributes. In their proposed approach, initially a tuple and then an<br />

attribute is selected for the insertion of watermark information. After selection of attribute value, a bit is<br />

removed from integer part of the target value and is added to float value of the same attribute. In this way<br />

distortion is produced in both integer and float type values.<br />

An obvious shortcoming with their technique is that it is applicable on the float type attributes only. Very<br />

few relational databases have float type attributes. . They are producing distortion to both integer as well<br />

float value of the attribute causing distortion at maximum level. Furthermore, data containing float values<br />

215


Sajid Iqbal et al.<br />

have more precision and can accommodate distortion to minimum level. They propose that their<br />

technique is resilient to secondary watermark addition attack but after the addition of secondary<br />

watermark attack to their watermarked relation, there is no way to find out which one of the watermark<br />

was embedded first.<br />

(Gupta 2006) proposed an algorithm for embedding watermark information to relational databases which<br />

is based upon the difference expansion of attributes values. Their technique selects two attributes from a<br />

single record and then on the basis of their difference, watermark information is embedded to selected<br />

attributes. Insertion of watermark is based upon the difference of attribute values.<br />

One problem is that they select two attributes from a single record which causes maximum distortion at<br />

record level. Furthermore, they claim that their technique is resilient to secondary watermark addition<br />

attack but after an addition attack, there is no way to find out which watermark information was<br />

embedded first to the relation?<br />

4. Proposed algorithm<br />

In this paper, we propose a novel algorithm to handle secondary watermark insertion attack. The main<br />

idea of the proposed algorithm is as follow:<br />

Database Owner calculates his watermark string by applying one way hash function which is SHA-1<br />

(SHA-1 NSA) in proposed case to his private key ko. He then sends his encrypted key to a Trusted Third<br />

Party which acts as a Time Stamping Authority. Time Stamping Authority (TSA) appends Trusted Time<br />

Stamp τ to the owner’s key and sends back the new key k to owner which is now owner’s key plus<br />

trusted time stamp. Then on the basis of trusted time stamp appended to owner’s key and percentage of<br />

watermark ω decided by the owner, a Unicode non-visible character Hair Space is embedded to<br />

calculated attribute for specific tuple.<br />

In case of an ownership conflict, owner can prove himself as real owner by regenerating his watermark<br />

information. Time stamp appended to the key proves that the watermark of claiming owner was<br />

embedded first in case of any secondary watermark attack.<br />

Our proposed algorithm is more robust as we are providing a two-level encryption security to the owner’s<br />

key. Besides a third party is involved at watermark insertion stage.<br />

4.1 Watermark construction<br />

Watermark information (trusted time stamp appended to owner’s key) is a series of binary string which is<br />

calculated by applying one-way hash function to owner’s key which is actually an image converted to<br />

binary string in our case. Then appending trusted time stamp to the key and again applying one-way<br />

hash function.<br />

Trusted time stamp is issued by trusted third party who acts as time stamping authority. Time stamp<br />

records the occurrence of a specific event at a specific time. Steps included in construction of watermark<br />

can be summarized as follow:<br />

Apply one-way hash function to owner’s key k which is private only to the owner.<br />

k = H (ko) where ko shows the owner key.<br />

Owner after applying hash function to his key sends it to Time Stamping Authority. They append Trusted<br />

Time Stamp to it and again apply a hash function.<br />

k = H (ko $ TTS)<br />

Once trusted time stamp is appended to owner’s key and hash function is applied upon it, the final key is<br />

sent back to the owner. The watermark bit string k which is now time stamp appended to owner’s key is<br />

embedded to the relational database to be watermarked.<br />

The watermarking procedure is shown in Fig 1:<br />

216


Figure 1: Watermark construction process<br />

4.2 Watermark insertion<br />

Sajid Iqbal et al.<br />

In contrast to current approaches, we propose a unique method for watermarking which does not cause<br />

any distortion at all. Instead of changing some least significant bits (LSB) which are mostly exposed to<br />

bit-flip attack and alteration attacks, we add a non-visible character space which is Unicode Hair Space<br />

to non-numeric attributes. Unicode hair space is a character space which is invisible, smaller than a<br />

usual space and does not produce any distortion at all.<br />

In order to distribute the watermark equally along the relation, the embedding process is cell-based. Up<br />

till now approaches have been adopted to embed watermark by first selecting a tuple and then specific<br />

attribute for that tuple.<br />

We are adopting cell based approach in order to distribute our watermark uniformly. The whole relation is<br />

considered as a one-dimension array of data elements and then watermark is embedded to them.<br />

A one way hash function is applied to owner’s defined key. The resulting bit string is sent to trusted third<br />

party, where a Trusted Time Stamp (TTS) is appended to the bit string and again a hash function is<br />

applied to the resulting string. The final key which is the owner’s defined key and trusted time stamp is<br />

returned to the owner.<br />

The embedding algorithm is described as: whereas table.1 shows list of notations used in algorithm<br />

217


Table: 1. Notations used in this paper<br />

Sajid Iqbal et al.<br />

Symbol Description<br />

Η Number of tuples in relation<br />

Ε Number of attributes in relation<br />

Γ Total Number of Cells in Relation (η * ε)<br />

Ω Percent of cells to be marked<br />

N No. of cells to be marked<br />

K Final Secret Key<br />

Τ Trusted Time Stamp<br />

Algorithm 1: Watermark insertion<br />

// Insert a watermark to relation R and return watermarked relation Rw<br />

// Parameters k, τ, ω all are private to the owner<br />

1. Calculate total Number of cells γ = η * ε<br />

2. Calculate number of cells to be marked N = γ * ω<br />

3. While marked_cells ≤ N<br />

i) if bit [H(ko concatenated τ)] = = 1<br />

ii) Add double space to Cell contents<br />

iii) Else add a single space<br />

iv) Increment bit stream of [H(k concatenated τ)]<br />

v) marked_cells = marked_cells + (γ / N)<br />

5. commit<br />

6. return Rw<br />

Some of the parameters including Key ko, time stamp τ and percent of cells ω to be marked are private to<br />

the owner and are decided by him.<br />

First of all in step 1 the total number of cells ‘γ’ in a relation is calculated. It is done by multiplying total<br />

number of tuples ‘η’ in a relation across total number of available attributes ‘ε’ to be marked. In step 2 we<br />

are calculating the number of cells N to be marked out of total available cells. The calculation is based<br />

upon the percentage of watermark as decided by owner. Once the calculation is done, the loop starts<br />

marking the specified cells according to the private key. If the bit of watermark string is binary ‘1’ then<br />

double Hair Spaces are added to the contents of cell under consideration and if the bit is binary ‘0’ then a<br />

single Hair Space is added to the contents of cell. Once all the required cells are watermarked, the<br />

operation is committed and a watermarked relation Rw is returned.<br />

4.3 Watermark detection<br />

If the owner of a database suspects piracy at some stage or there is a secondary watermark attack on<br />

his relation, he can prove his ownership by simply running watermark detect algorithm. Detection<br />

algorithm will successfully retrieve owner’s key and the trusted time stamp which can prove his real<br />

ownership showing that owner’s watermark was embedded first.<br />

218


Sajid Iqbal et al.<br />

As the watermark was embedded uniformly across the relation using cell based approach, detection<br />

algorithm will retrieve the whole pattern.<br />

The detection algorithm is described as follow:<br />

Algorithm 2: Watermark detect<br />

// Search relation R, extract time stamp τ and Key k<br />

// Parameters k, τ, γ all are private to the owner<br />

1. bit_stream = ‘’ //retrieved bit stream from relation<br />

2. Calculate total Number of cells γ = η * ε<br />

3. Calculate number of marked cells N = γ * ω<br />

4. Search relation for desired character<br />

5. While cells ≤ N<br />

i) if cell[contents] contains double space<br />

ii) append ‘1’ to bit_stream<br />

iii) else<br />

iv) cappend ‘0’ to bit_stream<br />

6. Return bit_stream<br />

In case of any piracy suspect or ownership conflict owner will run his watermark detection algorithm. First<br />

of all number of cells are calculated and then on the basis of percentage of watermark decided by owner<br />

at the time of insertion cells are searched for the desired pattern.<br />

On successful completion of detection algorithm, the results returned by detection algorithm are checked<br />

against a pre-defined threshold. If detected pattern reaches the threshold α ownership can be confirmed.<br />

5. Experiments and results<br />

The database that we used in our experimentation contains 14 attributes out of which 8 were numeric<br />

and 6 were non-numeric with one primary key attribute. 30,000 tuples were selected for experiment. We<br />

ran our algorithm on Intel dual core 2 GHz with 2 GB of RAM and running windows XP as operating<br />

system.<br />

5.1 Resistance against secondary watermark attack<br />

The basic purpose of our proposed algorithm is to handle Secondary Watermark attack. An attacker can<br />

add his factitious watermark as a secondary watermark to already marked database. In order to achieve<br />

this, attacker simply follows following steps:<br />

1. Take any factitious key Kf, apply some processing on it and prepare it as watermark string.<br />

2. Randomly select tuples and attributes for the addition of watermark string. Then randomly change<br />

the least significant bit, changing ‘1’ to ‘0’ and ‘0’ to ‘1’.<br />

3. And finally claiming the ownership of the database by retrieving his own fraudulently added<br />

watermark to prove his ownership.<br />

Since the attacker adds his factitious watermark after the embedding of original watermark, real<br />

ownership can be easily found. Whenever ownership conflict occurs, the real owner can retrieve his<br />

watermark which is actually his private key and Trusted Time Stamp representing the time at which the<br />

watermark string was provided and possible time of publication of database.<br />

219


Sajid Iqbal et al.<br />

In our experiments, a secondary watermark addition attack was simulated. A secondary watermark was<br />

added to a watermarked database, producing some distortion in attributes for randomly selected tuples<br />

with % of tuples selected for watermark was greater than that of actually embedded watermark.<br />

Then actual watermark detection algorithm was run. The detection rate of actual watermark was 99%.<br />

The retrieved string was the owner’s key and trusted time stamp which was the same as provided by<br />

trusted third party and same as date of publication of database. The retrieved string finally proved the<br />

ownership of real owner.<br />

5.2 Resistance against benign updates<br />

Along with handling secondary watermark attack, our proposed algorithm is highly resistant to subset<br />

attacks such as subset deletion, subset addition and subset selection attack. We simulated all these<br />

attacks against different ratio of ω = 90% and 50%. Although the proposed technique does not produce<br />

any distortion at all and hence can be applied to entire candidate records but we did not simulate such<br />

scenario.<br />

5.3 Subset selection attack<br />

In the subset selection attack, an attacker simply selects a part of the relation R. In our experiments we<br />

selected different number of tuples from relation R which was watermarked to two different extents i.e.<br />

with ω = 50% and ω = 90%. Experiments showed as shown in Fig. 2 that at both stages our watermark<br />

pattern was found in more than 75% of tuples which was quite enough for the claim of ownership. In the<br />

first go, 90% of total relation was watermarked and then the same relation was embedded with 50%.<br />

Figure 2: Detection rate on subset selection with ω = 90% and 50%<br />

5.4 Subset deletion attack<br />

In subset deletion attack, the copy right violator deletes a subset from the relation R in order to temper<br />

the imperceptible watermark and hence cause the ownership conflict. In our experiment we simulate the<br />

subset deletion attack by deleting at maximum of 90% of tuples and the watermark was still found at a<br />

very good ration as shown in Fig. 3.<br />

5.5 Subset alteration attack<br />

Subset alteration attack is little different in nature from subset deletion and subset selection. In this kind<br />

of attack, the attacker randomly selects attribute values for tuples and changes their least significant bits,<br />

i.e. flip 1 to 0 and 0 to 1. In our simulated experiments, our algorithm proved very robust against this kind<br />

220


Sajid Iqbal et al.<br />

of attack, as we are not producing any distortion in numeric fields. Simulated results against this kind of<br />

attack showed 100% detection of embedded watermark.<br />

Figure 3: Detection rate on subset deletion with ω = 90% and 50%<br />

6. Conclusion<br />

In this paper we propose a novel approach to handle the secondary watermark attack. Up till now,<br />

approaches have been used to embed a private key only. We embed an identification image as well as a<br />

trusted time stamp in our proposed scheme. Trusted time stamp can prove the ownership of real owner<br />

in case of any reversibility attack. Experimental results have shown that the proposed scheme is robust<br />

against different kinds of attacks. The proposed technique does not cause any distortion.<br />

References<br />

Agarwal, R., Kiernan, J., (2002). Watermarking relational databases. Proceedings of the 28th VLDB conference.<br />

China: pp. 155 – 166.<br />

Chen, X., Chen, P., He, Y., Li, L. (2008). A self-resilience digital image watermark based on relational<br />

database, Proceedings of the international symposium on knowledge acquisition and modeling. Wuhan: From<br />

21 to 22 Dec. pp. 698 – 702.<br />

Cui, H., Cui, X., Meng, M. (2008). A public key cryptography based algorithm for watermarking relational databases.<br />

In International conference on intelligent information hiding and multimedia signal processing. Harbin: From 15<br />

to 17 August. pp. 1344 – 1347.<br />

Dong, X., Li, X., Yu, G., Zheng, L. (2009). An algorithm resistive to invertibility attack. In watermarking relational<br />

databases, IEEE Control and decision conference. China: From 17 to 19 June. pp. 1587 – 1592.<br />

Gupta, G., Pieprzyk, J. (2009). Database relation watermarking resilient against secondary watermarking attacks.<br />

Proceedings of the 5th International <strong>Conference</strong> on Information Systems Security. India. pp. 222 – 236.<br />

Gupta, G., Pieprzyk, J. (2008). Reversible and blind database watermarking using difference expansion.<br />

“Proceedings of the 1st international conference on Forensic applications and techniques in<br />

telecommunications, information, and multimedia and workshop”. Australia.<br />

Hu, Z., Cao, Z., Sun, J. (2009). An image based algorithm for watermarking relational databases. In International<br />

conference on measuring technology and mechatronics automation. China: From 11 to 12 April. pp. 425 – 428.<br />

Lafaye, J. (2007). An analysis of database watermarking security, The third international symposium on information<br />

assurance and security. United Kingdom: From 29 to 31August. pp. 462 – 467.<br />

Li, Y., Swarup, V., Jajodia, S., (2005). Fingerprinting relational databases: schemes and specialties, IEEE<br />

transactions on dependable and secure computing 2(1). pp. 34 – 45.<br />

“The keyed-hash message authentication code (HMAC)”, http://www.csrc.nist.gov/publications/fips/fips198/fips-<br />

198a.pdf “Secure Hash Algorithm (SHA-1),” National Security Agency (NSA) – US.<br />

Wang, H., Cui, X., Cao, Z. (2008). A speech based algorithm for watermarking relational databases. In International<br />

symposiums on information processing. From 23 to 25 May pp. 603 – 606.<br />

X. Chen, P. Chen, Y. He, L. Li, “A Self-Resilience Digital image Watermark Based on Relational Database,” IEEE 2008<br />

International Symposium on Knowledge Acquisition and Modeling<br />

Zhang, Z., Jin, X., Wan, J. (2004). Watermarking relational database using image. Proceedings of the third<br />

international conference on machine learning and cybernetics. China: From 26 to 29 Aug pp. 1739 – 1744.<br />

Zhang, Y., Niu, X., Zhao, D., Li, J., Liu, S. (2006). Relational databases watermark technique based on content<br />

characteristic. In First international conference on innovative computing, information and control. Beijing: From<br />

30 th August to September 1 st . pp. 677 – 680.<br />

221


Engineering Change Through the Domains of Enterprise<br />

Architecture<br />

Tiko Iyamu<br />

Tshwane University of Technology, Pretoria, South Africa<br />

iyamut@tut.ac.za<br />

Abstract: Change continues to be a challenge, as organizations strive hard to change with time and with new and<br />

evolutionary processes, activities and requirements. Many organizations are thus challenged by this change. Some of<br />

these organizations attempt to address the challenges of change through different strategic approaches, such as<br />

Enterprise Architecture (EA), which involves the development of an organizational capacity to respond to changing<br />

needs and conditions. The research investigated the practice of EA to understand how EA, through its domains<br />

engineer Change in the organizations in order to respond to rapid technological and business changes. Interpretive<br />

case study approach, using two organizations was adopted.<br />

Keywords: change, enterprise architecture and domains architecture.<br />

1. Introduction<br />

Enterprise Architecture (EA) comprises of several different types of architectures, each with its own type<br />

of deliverables, analysis methods, processes and participants. The EA is defined differently by different<br />

experts, both academics and practitioners. The domains of EA include Business, Data, Information,<br />

Technology, Infrastructure, as well as Service Oriented Architecture. This research adopted the EA,<br />

which consist of four domains: Business Architecture (BA), Information Architecture (IA), Technical<br />

Architecture (TA) and Application Architecture (AA) (Cook 1996; and Zachman 1996). This is mainly<br />

because the two case studies adopted the same model of four domain approach. This helps the study to<br />

reflect and align on its objective. EA is a paradigm which premises to address challenges in the<br />

computing environment. According to the IT Architecture Affinity Group, “Architecture is a blueprint that is<br />

developed, implemented, maintained, and used to explain and guide how an organization’s IT and<br />

information management elements work together to efficiently accomplish the mission of the<br />

organization”. EA is said to enable change by bridging the gap between strategic planning and<br />

implementation efforts through a strategy process that is holistic in its coverage and enterprise-wide in its<br />

scope.<br />

On one hand, organizations’ activities including service delivery practice and organizational structure<br />

keep changing. On the other hand, information technology infrastructure also keeps changing and doing<br />

so rapidly. Nothing is permanent in organizations – not even for months. As a result, there has been<br />

emphasis on the significance of change in organizations (Finkelstein 1999). Many organizations are<br />

struggling to adapt to technological and business rapid change. This research investigated how EA,<br />

through its domains could be deployed to address fast changing need in the organizations.<br />

Where EA is adopted, the processes are a critical step in the growth and development of competitive<br />

advantage particularly in large organizations. This is due to the pace of change of activities, processes<br />

and requirements. According to Mack and Frey (2002), the primary reason for creating an overall IT<br />

architecture is to build the new as well as manage transition from the old. Change is not necessarily<br />

positive; it could be prohibitive and potentially pose risk to the organization. This is the primary motivation<br />

for the study, which led to the research question: What are the elements of change in EA and how<br />

could they be used to engineer change, thereby mitigating potential risk in the organization?<br />

EA is intended to provide the required flexibility to achieve change in the fast-paced, rapidly growing IT<br />

and business environments (Cook 1996; and Watson 2000). Iyamu (2009) argued that EA defines the<br />

scope, scale and nature of required changes within an organization and that it helps to identify the<br />

resources that must become involved. EA also provides a platform to consistently address all the<br />

activities in the organization and many related concerns such as the information and technology that<br />

supports the business processes and activities. This is the argument of some individual and group<br />

experts including Zachman (1996); Armour, et al (2007); Gartner; and The Open Group Architecture<br />

Framework (TOGAF). According to Lankhorst et al (2005), EA is a valuable instrument for<br />

operationalising and implementing policies and strategies.<br />

222


Tiko Iyamu<br />

EA is therefore inevitably seen as fundamental to strategic planning, systems design, software and<br />

hardware development, and the production of multiple systems with different functionalities but all from<br />

the same basic architectures (Youngs, et al 1999). As a result, irrespective of the requirements from both<br />

the computing environment and business, EA thereby ensures uniformity, reuse and manageability of<br />

change from the current state to realization of strategic intent, and thereafter.<br />

In many of the organizations, EA serves as a tool for identifying and driving reuse throughout the existing<br />

and planned architecture. This is not as smooth as it is claimed. According to Spewak (1992), from the<br />

development to implementation stages of EA, there are challenges from both technical and non-technical<br />

perspectives. For other organizations, EA serves as an agent in managing application and technology<br />

infrastructure change within information systems supporting business transactions and information<br />

processing (Zachman 1999).<br />

Due to the complexity of the study, appropriate research method was key, in order to achieve the<br />

objective. This was to ensure a comprehensiveness and richness of the data. The adopted method is<br />

described in the next section.<br />

2. Research approach<br />

The study employed a qualitative, case study approach to investigate the elements which engineer<br />

Change in EA. It is argued that the qualitative methodology is more suitable for this type of study as it<br />

allows clarification of questions. According to Walsham (1995), qualitative research is a very useful<br />

method for complex situations and theories. It has been argued that case study materials are drawn from<br />

different means including interview questions, structured and semi-structured interviews with some of the<br />

key players, documentary sources and ad hoc observational and experience-based notes (Lee 1989; and<br />

Yin 1994). The case study method was applied primarily because it allows in-depth exploration of the<br />

complex issues involved in this study. Two case studies were conducted with a Financial institution and a<br />

Government institution. The two organizations were selected on the basis of prima facie evidence that<br />

each provides a good example of an organization subject to Change as engineered by EA, and each<br />

provides some evidence of success and failure in the process. The two case studies were adopted to<br />

gain an insightful, qualitative interpretation of the changes which were happening as a result of the<br />

practice of EA in the organizations.<br />

The Financial institution is referred to as “Company A” and “Company B” the Government institution.<br />

There were 150 and 80 employees in the computing environments of Company A and Company B,<br />

respectively. Each of the companies has been in operations for more than hundred years.<br />

Data sources included interviews and documentation. As shown in Table 1 below, the number of<br />

interviewees varied, depending on the size of the organizations. A set of balanced respondent<br />

demographics was a key factor in achieving a true reflection of the situations. Targeted respondents<br />

were from different units and were at various levels of the organizational structure within the Business<br />

and IT departments. They included Business and Systems Analysts, IT Architects, IT Managers, IT<br />

Project Managers, IT Executives and Business Managers. The interviewees, as tabulated in Table 1<br />

below, were selected on the basis of their closeness to the topics of EA, including their individual levels of<br />

experience in management, organizational and architectural issues. In the study, closeness is defined as<br />

the area of specialization, with significant amount (5 years) of stock of knowledge and experience. In<br />

Company A, the Chief Technology Officer (CTO) and in Company B, a senior manager assisted in<br />

identifying the interviewees in their various organizations. Employees also assisted in identifying relevant<br />

interviewees.<br />

Structured and semi-structured interviews were designed to elicit information on what the Change<br />

elements were in the practice of EA. Questions were asked around the understanding of EA; the<br />

perception of the impact and implication resulting from the practice of EA in the organization; and<br />

involvement and levels of commitment of actors in the organizations.<br />

The data was analyzed using the interpretive approach. Interpretative rules create signification or<br />

meaningful symbolic systems that provide ways for actors to see and interpret events. Agents reflexively<br />

apply interpretative schemes and stocks of knowledge. The interpretive approach looks at ‘reality’ from a<br />

different perspective (Walsham 1995). According to Kaplan and Maxwell (1994), in an interpretive<br />

research project, there are no predefined dependent and independent variables, but a focus on the<br />

complexity of human sense-making as the situation emerges.<br />

223


3. Analysis of the case studies<br />

Tiko Iyamu<br />

The analysis was done to understand how EA engineer change in the organizations. The results of the<br />

case studies are combined and presented in this section.<br />

As mentioned in the introduction section, both case studies adopted the model as advocated and<br />

supported by consultants and groups including Zachman (1996); Gartner; and TOGAF. The model<br />

categorizes domains of EA: Business Architecture (BA), Information Architecture (IA), Technical<br />

Architecture (TA) and Application Architecture (AA) were the domains which constituted EA in the<br />

organizations. Each of these architectures has a specific set of deliverables, analysis methods,<br />

processes and participants through which they facilitate change within the computing environment and<br />

between IT and Business departments. Thus EA has made ongoing representation of the existing and<br />

planned information systems required to support the ever-changing business strategies. This included<br />

the consequential business processes, information, and application requirements and technology<br />

selections.<br />

All the activities and processes which were happening in the computing environment of the case studies<br />

at the time were categorised into domains of BA, IA, TA and AA. This includes selection, deployment and<br />

use of technologies as well as the services within the organizations and their clients. There were many<br />

architects in each domain. Each domain was managed by a senior architect, titled architecture manager<br />

and team leader in companies A and B, respectively. The EA was managed by Chief Architect and<br />

General Manager in Company A and Company B, respectively.<br />

Each domain had its guiding Principles, Standards and Policies through which their various mandates<br />

were implemented and managed. These principles, standards, policies were defined based on IT and<br />

business requirements and interpreted by the architects. As a result, the roles of the architects were<br />

critical in the deployment of EA in the organizations. As such, not anyone could be appointed to take on<br />

the task of architect, it required special skill. For example, in Company A, the evaluation and<br />

recommendation of server virtualization and consolidation which was conducted in the organization<br />

required in-depth knowledge of specialized skill. The project was led by the architecture unit on behalf of<br />

the organization.<br />

One of the main requirements for the deployment of EA in both organizations was to bridge the gap<br />

between the business and technology components of the organization. This requirement was<br />

emphasized and prioritized in the organizations. This occurs during the process of development and<br />

implementation of EA. In order to achieve the required flexibility to handle the fast-pace of IT, change<br />

was inevitable. This was a serious challenge in both case studies. As a result, there was emphasis on<br />

business vision and requirements, how they were extracted, documented and achieved.<br />

Technical requirements were as important and critical as the business requirements in engineering<br />

change through EA. Technical requirements were defined, implemented in both technical and application<br />

domains of EA. The drivers were however derived from information and business architectures. Some of<br />

the business managers acknowledge the contribution and role of EA in their various processes. A<br />

business manager in “Company A” had appreciatively expressed as follows: “Shareholder Value would<br />

not be successfully achieved if IT didn’t support and influence aligned business goals – imagine the<br />

business driving client centricity and IT ignoring this and continuing to push silo- or product-centric<br />

solutions, thus jeopardizing or even failing to drive projects to focus on the client (like Client Relationship<br />

Management), and thus continue to compete for the client’s business and not integrating IT solutions<br />

across the organization for a single client experience and view; such unaligned behaviour is of the past<br />

and cannot drive profitability, as the client inevitably gets annoyed with the corporate organization for<br />

wasting his time and hence taking his business elsewhere”.<br />

However, available skill-set posed a serious challenge, especially in the area of business architecture.<br />

Company B didn’t have enough skilled people to carry out the tasks as dictated by EA. The other<br />

organization fared better. According to one of the interviewees, a senior manager in Company B, the cost<br />

for the retention of people with specialized skill was prohibitive. The manager admitted that in some<br />

areas, change couldn’t be implemented as a result of lack of technical know-how.<br />

According to an interviewee, a senior manager in Company A, one of the biggest challenges in managing<br />

change is gaining an accurate view of the current state. It was clear, EA is critical and significant in<br />

224


Tiko Iyamu<br />

providing a more accurate view of the current quo. Once the status quo was identified, a plan of action<br />

followed. EA further facilitates the assessment of the impact of Change on the current business<br />

environment and the conception of strategic alternatives for consideration. According to one of the<br />

interviewees, “Yes definitely, the company is now able to consolidate technologies and business<br />

solutions, as well as more rapidly engage in mergers and acquisitions – architecture is an intellectual<br />

capital investment increasing organizational maturity from tactical to strategic”.<br />

The organizations were determined to establish current-state of the environment in order to understand<br />

the gap, if any. Particularly, senior executives were keen to understand how EA could be applied to<br />

reasonably articulate their future business environments and resulting strategies and then use gap<br />

analysis to assess the impact of change and associated risks. To achieve this, the cooperation of both<br />

the computing environment and the business units of the companies were very important, including the<br />

support of the executives and other managers in the business. The IT architects realized and admitted<br />

the criticality of the support and buy-in of these managers to the success or failure of EA in the<br />

organization. Interviewee, some of them, architects in Company “A” explained as follows: “More and<br />

more business managers walk the architecture talk. I’ve often been in discussions between business<br />

managers and outsiders, where the business manager’s first question would be how proposed solutions<br />

fit with our defined architecture”. Similarly, in Company “B” stakeholders understood the importance of<br />

business buy-in. One of the interviewees, an architecture manager says: “Get the business buy-in to the<br />

architecture process first, and let them sell it to the rest of the organization. . . A closer alignment with<br />

business was formed in order to ensure a speedy and successful implementation”.<br />

The scope of EA was therefore the union of the organization, the business reengineering and<br />

development that was applied to it, and the technical domains that supported it. Thus, EA was defined<br />

through a pragmatic need: the need to design and redesign as well as continuously improve the<br />

functioning of organizations. EA can be applied in different environments and contexts to facilitate<br />

change. A business manager in “Company B” says “the business architecture has helped in the design,<br />

documentation and implementation of process-oriented activities in the organization”. This view was<br />

echoed in many quarters of the organization. Processes were easier to follow as a result of avoidance of<br />

re-inventing the wheels. Processes were reused in accordance to set rules and definition of the various<br />

domains of the EA.<br />

In the organizations used in the study, change happened in the areas of process, information flow,<br />

business function, technology innovation, as well as personnel movement. The frequency of change was<br />

a huge challenge, and the impacts and implications were of significant importance to both organizations.<br />

The changes were instigated and influenced by internal and external factors. The internal factors<br />

included business as well as IT vision, strategies, roles and responsibilities, technical know-how and<br />

budget. While the external instigating and influencing factors were technology innovation, industry trends,<br />

market competitiveness and government policy.<br />

Each component of architecture provided a unique view of the enterprise leading to its own capabilities,<br />

primarily: the BA provided the tools, models, techniques, and participants to manage the impact of<br />

change on business processes, clients and partners. Similarly, the IA enabled the management of<br />

change on information flow and exchange between applications as well as processes; the TA enabled<br />

the management of change on technology infrastructure; and the AA enabled the development and<br />

portfolio management of business application and software.<br />

However, the level of success or failure was not measured as it was out of the scope of the study. This<br />

could be an opportunity for future research work.<br />

The study revealed that EA was typically not mastery of a single innovation, but ongoing learning,<br />

adaptiveness and the development of collaborative work cultures. It was admitted that changing the<br />

culture of the organization was the highest ambition. According to one of the managers in “Company A”,<br />

“EA seem to be helping in achieving the challenges of rapid change in the organization”. Even at the<br />

demanding level, EA helped by giving all concerned a shared view of how the business works, and what<br />

it works within the context of the organization.<br />

The organizations’ strategies were the means and the domains of EA were the ends in an attempt to<br />

achieve the vision of the organizations. The study also revealed that the change elements of EA impact<br />

both technical and business components of the organizations. This was a key factor in bridging the gap<br />

225


Tiko Iyamu<br />

between the business and IT through the domains of EA. The change elements in EA had strong<br />

influence on the development and deployment of technology strategy and the amalgamation of both the<br />

IT and the business units towards one vision.<br />

Each domain had its specific template, which was designed to help achieve both business and IT<br />

objectives. The template was used to review the current state, mainly to detect deficiency, identify gap<br />

and plan for the future. The review was a major process carried out every quarter in Company A. In<br />

Company B, it was an exercise, which was done only in the technical domain on yearly basis.<br />

The templates made prediction and specifics possible in the various domains. The areas covered by the<br />

domains became stabled, uniformed and easily accessible. According to some employees in the two<br />

organizations used in the study, the templates reduced guesswork and try-and-error. Thus, encouraged<br />

and helped bridged the relationship gap between architects and other employees in the IT department,<br />

on one hand, and between the IT and business departments on another hand.<br />

4. Findings<br />

The above analysis reveals some findings. The most critical of them are presented below. The headings<br />

were reached based on what were common and critical in both organizations:<br />

4.1 Business requirements<br />

The business requirements were extracted from the companies’ strategies. They included operational<br />

goals and strategic intent, which were being pursued at the time. It focused mainly on the strategic<br />

objectives of the organization. The requirements included the drivers and trends of the organizations’<br />

vision and strategy. These drivers were influenced by factors such as time, economics and<br />

environmental trends.<br />

In order to achieve the required flexibility to handle change in the fast-paced, rapidly growing IT and<br />

business environments, EA was deployed with the intention of addressing the organizational needs and<br />

to facilitate change through the adaptation of innovations of technology to business process model, in a<br />

uniform manner. This was said to be successful only if it aligns with the business strategy and have the<br />

buy-in of the business executives.<br />

4.2 Technical requirements<br />

The technical requirements were based on the business processes and activities, which were meant to<br />

support and enable competitive advantage. The requirements drive the introduction and selection of new<br />

technology by focussing on function and scalability, compatibility and interoperability. It translated the<br />

business requirements into common terminology, which enabled consistent semantic meaning across<br />

information systems and technology in the organizations. These were the primary focus of the TA and AA<br />

domains. Through this approach, for example, technologies were certified usable or obsolete in terms of<br />

manageability. The change in technologies often had impact on business processes.<br />

4.3 Process oriented<br />

The analysis revealed that EA is a process oriented approach, not an event or project. This seemed to<br />

have been understood by many of the promoters in the organizations. For example, it required continued<br />

business and technical requirements to be translated and technology selected to respond to changing<br />

needs. It continually facilitated other activities such as concept reuse, system role definition, and the<br />

modeling of optimal information flows. It guided organizational learning and constantly enabled learning<br />

of new things and subsequent sharing of that learning. One of the examples as revealed in the study<br />

was, if the direction of the Enterprise’s vision changes, then resources and infrastructure has to be<br />

managed so that they could be shared in order to achieve the vision. This was done in a virtual circle<br />

manner and it became the norm in Company A.<br />

4.4 Skill set<br />

In EA, it was argued to have adequate specialized skill set as the execution of its functions required more<br />

than basic understanding hence skill set was critical. However in both case studies this was a challenge.<br />

Specialized skill sets were required in all the above listed findings. The issue of skill was as important as<br />

the rest of the factors in the development, as well as in the implementation of EA to engineer change in<br />

226


Tiko Iyamu<br />

the organizations. Change requires the development of an organizational capacity to respond to<br />

changing needs and conditions within scope and periodically. This is only achievable through skilled and<br />

adaptability of personnel, including the architects. The architects engineered the process of change<br />

through domains of EA, which was under their auspices. The organizations fell short in some areas of the<br />

domains, such as business.<br />

Change involves adaptations in practices, skills, and often beliefs about what is important and valuable,<br />

such as requirements of both technical and business. It required the interaction and connectedness to<br />

ensure uniformity, alignment of skills, business and technology processes. In addition, change was<br />

intended to enable sharing of power across different components of the organizations. These were very<br />

significant and were only achievable through specialized skill set.<br />

5. Interpretation of the findings<br />

The above four findings were interpreted and six critical factors were extracted. The factors are<br />

Adaptiveness, Innovation and Governance, Culture, Uniformity, Strategic Alignment and Technical Knowhow.<br />

These factors were primarily, the taproots in engineering change through EA. They were critical and<br />

prevalence in the activities of the EA domains.<br />

5.1 Adaptiveness<br />

EA is not a "quick fix" solution for the replacement of lack of planning. It requires adaptiveness. This<br />

enables it to effectively and efficiently build, maintain, and apply the entire domains of EA. It is more<br />

beneficial when it is applied uniquely to each need of the organization, depending on its business<br />

strategies, architectural maturity, priorities, corporate culture, and political environment. Its adaptiveness<br />

enabled any of the domains to address the potential risk as well as enable and support the movement<br />

from current to the desired state. Change must be adaptive to achieve its goals, either in business or<br />

technology.<br />

5.2 Innovation and governance<br />

Through its change process, EA attempts to find different and better ideas and ways of enhancing and<br />

increasing the companies’ productivity in a competitive market. Innovation was well guided, this including<br />

the selection and deployment of technology and process engineering. Through domain, EA provides<br />

governance in many activities such as data usage, software development and infrastructure deployment<br />

and management in the computing environment of the companies. For example, in Company B, EA was<br />

used to carry out investigations and implementation of server consolidation and virtualization as dictated<br />

by the requirements of the organization.<br />

5.3 Culture<br />

The rules, norms and procedures, together, created a specific culture, which were institutionalized over<br />

time in the individual company. According to Iyamu (2009), institutionalization is the process where a<br />

practice is assimilated into the norm. It is not easily disassociated, dismantled or re-designed. This<br />

affected the human actors and was also reflected in the use of technology. For example, no software<br />

could be installed on the network server without a written approval from the domain expert in architecture<br />

team. This was the norm. It helped to change individual’s attitude and approaches in carrying task in the<br />

computing environment.<br />

Culture was one of the key elements through which EA impacted change in the organizations. Within the<br />

scope of the organizational requirements, EA was developed and implemented and practiced. Certain<br />

factors of a personal nature need a particular organizational culture in which to thrive, or would feed on<br />

particular internal policies. Similarly, certain internal policies would only be possible within a particular<br />

organizational culture.<br />

5.4 Uniformity<br />

EA provided a uniform process for selecting technology infrastructures, of documenting the current state<br />

of business process, data and infrastructure; future states; and the gap between the current and future<br />

states. This applies to the autonomous business units of the organizations. This was critical to providing<br />

support for both the business processes and technologies in terms of factors such as technological<br />

enablement reuse, licensing optimization and skill management.<br />

227


5.5 Strategic alignment<br />

Tiko Iyamu<br />

The EA will not work if there is no understanding between the business and IT specialists and<br />

executives, but that it can encourage that understanding. The understanding typically concerns change<br />

of processes, management, methods, technology infrastructure design, pattern and selection. EA was<br />

instrumental in facilitating closer relationships as well as bridging the gap between the business and IT<br />

through sharing of a common view and vision of what had to be done and how. Similarly, each domain<br />

was aligned to the IT and business strategies of the organization.<br />

5.6 Uniformity<br />

The EA approach facilitated processes, activities and procedures within the computing environment of<br />

the organizations. This included all the domains of EA and began from project initiation to implementation<br />

stages. It facilitated framework to manage and share information and to ensure that the business<br />

processes and activities were supported by technology infrastructure and applications that provided the<br />

needed data, information flow and the infrastructure to enable the data.<br />

The EA facilitated change, which affected not only business processes, clients and partners, but also<br />

information, supporting business applications, and the technology infrastructure. This made it a<br />

challenging task to attempt to manage change by any means, including the adoption of EA. The evidence<br />

from the case studies revealed the elements, which made EA a change agent. On the basis that the<br />

change was the focal factor for EA deployment, these elements were also used as measurement criteria<br />

for success and failure of EA in the organizations.<br />

6. How EA engineer change through its domains<br />

The findings and the interpretation presented above are now been mapped to the domains of EA, from<br />

the perspective of change. The mapping was done to make sense of the interrelationship between the<br />

findings and the interpretation and how they engineered change through the domains of EA.<br />

As identified in the findings, the most critical factors required in engineering change through the domains<br />

of EA were Business Requirements, Technical Requirements, Process Oriented and Skill Set. This is<br />

illustrated in Figure 1 below. Based on these factors, the EA formulated principles, standards and policies<br />

within the various domains. These domains engineered change through identified elements. The primary<br />

aim was to move the organization from the current to future state.<br />

Resource EA Domain Change Element<br />

Business Requirements Business<br />

Architecture<br />

Technical<br />

Requirements<br />

Information<br />

Architecture<br />

Process Oriented Technical<br />

Architecture<br />

Skill-Set Application<br />

Architecture<br />

Adaptive<br />

Innovation &<br />

Governance<br />

Culture<br />

Uniformity<br />

Strategic Alignment<br />

Technical Know-how<br />

Figure 1: Engineering change through EA domains<br />

The domains were not independent of each other in the business processes and technical activities of<br />

the organizations. Change to one domain had impact on others. The domains collaborated with each<br />

other. This is illustrated with the up and down link arrows in Figure 1. Typically, businesses pursuits<br />

improved or reacted to change schematically.<br />

228


Tiko Iyamu<br />

The discussion that follows should be read together with Figure 1 to get a full appreciation of how<br />

elements of resources constituted themselves into the domains, through which change was engineered<br />

in the organizations.<br />

The next section provides how EA engineered changed through its domain in the organizations. The<br />

change effort is articulated on generalization. Generalization is an explanations of particular phenomena<br />

derived from the empirical interpretive study in the specific settings, which could be valuable in the future<br />

in other organizations and contexts.<br />

6.1 Business architecture<br />

Base on the organizations’ strategy, business requirements were formulated. The BA developed<br />

principles, standards and policies to provide guidance for the development of response plan. It also<br />

ensures uniformity of processes, where possible, across the organizations. In addition, the BA:<br />

Led the executive team and the line-of-business managers through the development and usage of<br />

their various strategic models, which was intended to be adaptive to changing needs.<br />

The gap between strategy development and tactical decision-making was a shortcoming of many<br />

strategic plans. BA eliminated much of the guesswork for operational managers in interpreting the<br />

impact of strategic plans on operations, and the priority that was placed on different plans. This was<br />

done by providing guidance concerning the organization information assets to knowledge workers,<br />

information processors, IT application developers, infrastructure managers, and executives.<br />

6.2 Information architecture<br />

The elements were able to effect Change mainly through their impact on the organization’s key factors.<br />

The IA defined availability and usage of information for business requirements and other key business<br />

processes and enabled the underlying processes and skill-sets for changing needs in the organization. IA<br />

primarily engineered change by:<br />

Defining the sources of information and ensure the availability and usage of this information by the<br />

key business processes and enabled by the underlying application and technical architectures.<br />

Identifying the information flows that was intended for optimization (increased velocity, density, and<br />

reach) as well as the information entities that had to be defined and used consistently across the<br />

information value chain to increase the value of information across the organizations and the external<br />

transactions.<br />

6.3 Technical architecture<br />

In the deployment of new technology, based on functional and utility advantages, the TA facilitated<br />

alignment between business and technical requirements and ensures compatibility and interoperability<br />

with the strategic infrastructure. This was a key to rapid response continual changes in the organization.<br />

It provided configuration standards and guidelines for using standard products in the environment. TA<br />

was adopted to manage technology change in two primary ways:<br />

Provided the principles, standards, configurations, process, and governance mechanisms to<br />

establish and maintain an adaptive technology infrastructure and application reuse methods.<br />

Drove the introduction of new technology based not only on functional and utility advantages, but<br />

also on compatibility and interoperability with the existing architecture. This was key to rapid<br />

implementation of infrastructure configuration standards; guidelines for using standard products; and<br />

principles for integration. These provided both a view of the recommended technology and basis for<br />

assessing the impact of new and replacement technologies within the context of the whole<br />

organization.<br />

6.4 Application architecture<br />

The application architecture defined the way applications were designed and deployed, and how they<br />

cooperated with each other. The AA provided standards and guidelines that supported Build, Reuse and<br />

229


Tiko Iyamu<br />

Buy of robust application for business productivity gains. The AA employed and managed application<br />

change through oriented process, innovation and adaptiveness. Other focus areas included:<br />

Defined the gap between the existing systems functionality and the functionality needed to satisfy BA<br />

and IA requirements.<br />

Provided a framework for migration by taking into account not only functional requirements, but also<br />

investment strategy in information systems and the “technology fit” of existing and planned systems<br />

with the TA and technology infrastructures.<br />

Responsible for reuse, this was a key component to success in the drive to facilitate change, by<br />

guiding developers toward systemic analysis, design and development of business components and<br />

objects.<br />

In summary, EA provides standards based planning, development and implementation methodology that<br />

will help IT strive to meet the rapid changes in the business processes. It allowed a translation of<br />

functional requirements to the selection of services, standards, components, configurations, their<br />

phasing, and the acquisition of products that implement them – much more than the project management<br />

and systems analysis disciplines that have been widely adopted over the last thirty years in trying to<br />

solve the same problems. Some independent consultants have in the past attempted to generically<br />

address some of these elements of Change. This has however been without empirical evidence as in this<br />

study.<br />

7. Conclusion<br />

Neither Change nor the development and implementation are simple in any organization. The return on<br />

investment depends on how well business and technical requirements are articulated, and how the<br />

people interpret the process and the technology they are using to support and enable the business. The<br />

contribution of the study arises from implications for the decision makers responsible for sponsorship of<br />

EA, including the architects. These decision makers need to understand the functions, dynamics and<br />

causes of what, why and how EA bring about Change on both business and technology, and how the<br />

impact can be ensured and measured. The other contribution of this study aims to be of significance to<br />

decision makers, professionals, including managers and employees of the organization within the<br />

computing environment, and IS researchers. It is expected that the key contribution will arise from the<br />

understanding of the fundamental elements through which EA engineer Change. Through this, a better<br />

understanding of the contribution of non-technical factors in the deployment of EA is gained.<br />

References<br />

Armour, F., Kaisler, S. & Bitner, J. (2007): Enterprise Architecture: Challenges and Implementations. HICSS, 40th<br />

Annual Hawaii International <strong>Conference</strong>, System Sciences, volume, Issue, pp. 217 – 217.<br />

Cook, M.A. (1996): Building enterprise information architectures: reengineering information systems, Prentice-Hall,<br />

Inc., Upper Saddle River, NJ.<br />

Finkelstein, C. (1999): The Pace of Change http://members.ozemail.com.au/~visible/papers/dmr02.htm<br />

Gartner. The Business of IT. Accessed 13 February 2008. http://www.gartner.com/technology/research.jsp<br />

Iyamu, T. (2009). The Factors affecting Institutionalisation of Enterprise Architecture in the Organization, Proc.<br />

<strong>Conference</strong> on Commerce and Enterprise Computing (CEC09), IEEE Computer Society, July 2009, pp. 221-<br />

225.<br />

Kaplan, B. & Maxwell, J. A. (1994). Qualitative Research Methods for Evaluating Computer Information Systems. In:<br />

J.G. Anderson, C.E.<br />

Lankhorst et al. (2005): “Enterprise Architecture at Work: Modelling, Communication and Analysis”, Springer.<br />

Lee, A.S. (1989): A scientific methodology for MIS case studies, MIS Quarterly, vol. 13, no. 1, pp. 33 – 50.<br />

Mack, R and Frey, N. (2002). Six Building Blocks for Creating Real IT Strategies. Gartner Group Inc.<br />

Spewak, S.H. (1992): Enterprise Architecture Planning: Developing a Blueprint for Data, Applications and<br />

Technology, John Wiley & Sons Inc., New York.<br />

The Open Group Architecture Framework (TOGAF). Understanding the IT architect's place in today's business<br />

environment. Accessed 27 June 2008. http://www.ibm.com/developerworks/library/ar-togaf1/.<br />

Youngs, R., Redmond-Pyle, D., Spaas, P. & Kahan, E. (1999): A standard for architecture description, vol. 38, no. 1,<br />

Enterprise Solutions Structure.<br />

Yin, R. K. (1994): Case Study Research, Design and Methods, 2 nd ed., California, Newbury Park; Sage Publications.<br />

Walsham, G. (1995). “Interpretative Case Studies in IS Research: Nature and Method”. <strong>European</strong> Journal of<br />

Information Systems, vol. 4, no. 2, pp. 4–81.<br />

Watson, R.W. (2000): An Enterprise Information Architecture: A Case Study for Decentralized Organizations. HICSS,<br />

p. 7059, 33rd Hawaii International <strong>Conference</strong>, System Sciences, vol. 7.<br />

Zachman, J.A. (1996): Enterprise Architecture: The View Beyond 2000, <strong>Conference</strong> Proceedings, Warehouse<br />

Repository Architecture Development 7th International Users Group <strong>Conference</strong>, Technology Transfer Institute.<br />

230


Evaluating Interoperability Artifacts for the Exchange of<br />

Public Service Information: Outline of a Conceptual<br />

Framework<br />

Veit Jahns<br />

University of Duisburg-Essen, Essen, Germany<br />

veit.jahns@icb.uni-due.de<br />

Abstract: Drivers to improve the interoperability of information systems in public authorities are initiatives for<br />

simplifying the access to public authorities and their services for citizens and enterprises. The spectrum of such<br />

initiatives reaches from initiatives on the regional and national level up to global initiatives. All these initiatives have in<br />

common, that they are related in some way or the other to public services. I.e., they have to deal with the challenge<br />

of describing public services either as a base for exchange information about these public services within the public<br />

administration or between public authorities and citizens and enterprises or as a base for developing information<br />

systems to support the provision of public services. As a result of these initiatives numerous artifacts have been developed,<br />

including languages and methods for modeling public services, ontologies and taxonomies for public<br />

services, etc. Although, this multitude of artifacts has the positive effect, that there is a high chance, that every aspect<br />

of the complex concept “public service” is covered by at least one of these artifacts, it is difficult to keep track on all<br />

these artifacts and how are these artifacts are related to each other, in particular what interoperability issue they<br />

address in detail, on which foundations and assumption these artifacts are based on, etc. In this paper a conceptual<br />

framework for evaluating such artifacts is proposed, which shall allow an evaluation of these artifacts with respect to<br />

the question given above. But this conceptual framework can be also useful for support of an artifact to solve a<br />

particular description problem regarding public services on the one hand, and on the other hand, it can be useful for<br />

the integration of these artifacts.<br />

Keywords: eGovernment, evaluation, public service, interoperability, artifact<br />

1. Introduction<br />

Once Garfield (1977: 175) with the numerous scientific journals and the need for literature reviews in<br />

mind stated, that maybe there are too many journals. The same can be stated for interoperability artifacts<br />

today: there are (too) many of them. Driven by private and public initiatives standards, models, as well as<br />

modeling languages and methods have been developed to address different interoperability problems<br />

and issues. This applies not only for the whole field of interoperability between information systems in<br />

public and private organizations, but also in the field of interoperability of repositories for public services.<br />

The objectives of such initiatives with respect to public service repositories are to simplify the access to<br />

public services as well as to improve the production of public services by a better integration of public<br />

authorities across administrative as well as geographical boundaries. Examples for such initiatives are<br />

projects within the <strong>European</strong> Union for the implementation of the <strong>European</strong> Service Directive, which demands<br />

a single point of contact, where a company can handles all its concerns with the public administration.<br />

But this doesn't require, that the single point of contact has to provide all these public service,<br />

but rather that this single point of contact manages the communication and transaction with the<br />

authorities providing the respective public services, and thus, needs to be interoperable regarding these<br />

public services with the respective, in particular with the respect to the exchange of information about<br />

public services.<br />

In addition, the objective of simplifying the access to public services gains more relevance by developments<br />

on the political level. E.g., it is stated in the “Ministerial Declaration on eGovernment” by the<br />

<strong>European</strong> ministers responsible for the eGovernment policy in their countries (<strong>European</strong> Union 2009),<br />

that it is the common aim of the respective states to facilitate and simplify the access to public services –<br />

not only for their own citizens and enterprises, but also for the citizens and enterprises from the other<br />

states. These efforts finds expression in many research and implementation projects that have been<br />

started in the past, e.g., TERREGOV (http://www.terregov.eupm.net), Access eGov (http://www.accessegov.org)<br />

or SemanticGov (http://www.semantic-gov.org), or are underway, e.g., Peppol<br />

(http://www.peppol.eu).<br />

Besides this Euro-centric examples many other examples for such initiatives dealing in one way or the<br />

other with interoperability issues regarding the provision of public services can be found in America, Asia-<br />

Pacific or even on the global-level. But to shorten the discussion, in all these projects, initiatives and<br />

efforts many different artifacts were created: ontologies, data models, modeling and markup languages,<br />

231


Veit Jahns<br />

etc. In general, these artifacts serve the purpose to describe public service in a unified manner, but every<br />

project has a particular perspective on public services. So this rise the questions, if they address the<br />

same interoperability issues or do they focus only on selected and distinguished interoperability issues, if<br />

these are artifacts based on the same technological and methodological basis, and how these artifacts<br />

are related to each other? Is there a reason, why there are so many different interoperability artifacts or is<br />

this field—as Frank (2000: 48) once self-critically states—just another play-ground for researchers and<br />

the purpose of the artifacts’ creation just to satisfy its creators (Frank 2000: 48)?<br />

Against this background it seems to be a promising task to review and evaluate the current state-of-the<br />

art of interoperability artifacts for modeling and describing public services, in particular to answer<br />

questions like how are these artifacts related to each other, i.e., do they compete or complement each<br />

other, which interoperability issues are addressed by the interoperability artifact, what are the technological<br />

and conceptual foundations of these artifacts, what assumptions about the domain of public<br />

services are made for the developing of these artifacts, etc. Answers to these questions will be useful to<br />

guide the further development of interoperability artifacts To this end a conceptual framework to support<br />

such a review and evaluation of interoperability artifacts will be developed and presented in this paper.<br />

The paper is organized as follows: after the introduction the two key concepts of this paper public service<br />

information and interoperability artifact will be discussed and defined. The third section is dedicated to the<br />

elaboration and justification of the conceptual framework. A conclusion with future research questions<br />

finishes this paper.<br />

2. Definition of key concepts<br />

In this paper two key concept are used, and therefore have to be defined: public service information and<br />

interoperability artifact. The concept public service information shall be defined within this paper as<br />

information about public services. In the first place, this concept seems to be clear with this definition, but<br />

looking at this concept a little bit closer this concept will be become vaguer. Then what means public<br />

service and what is information about public services? The first question is difficult to answer, because<br />

what a public service is and shall be depends on the weltanschauung by the person defining the concept<br />

of public service. Thus, providing a definition, which is acceptable for everyone (Jäckering 1991: 18–19)<br />

is a difficult task, if not impossible. Hence and for pragmatic reason, public service will be defined very<br />

generally and with the aid of Lynn (2006: 24) as a service provided by an organization assigned to it by a<br />

sovereign to provide this service. And public service information will be defined as information about this<br />

service that give indication who provides this service, how this service can be accessed, which<br />

requirements have to be fulfill to access this service and so on. But as a consequence of such a definition<br />

of the concept of public service, the manifestations of public services cover a wide spectrum.<br />

Furthermore, this spectrum possesses a high dynamic, i.e., services sometime lose their characteristic to<br />

be a public service and will be provide by private organizations, or gain it, i.e., will be provide by a public<br />

organization instead of a private organization (see e.g. Fugge 2003).<br />

The second key concept of interoperability artifact has been used before by Fustar (2009: 1–2) in his<br />

Interoperability Meaning Diagram. In this diagram he relates the concept of interoperability artifact to<br />

concepts like standard semantic model, reference model, components and so on. As the discussion in<br />

the introduction indicates, a semantic model can be also seen as an artifact: a man-made tool to address<br />

a particular interoperability issue. Furthermore, Fustar does not give a detailed definition, what an<br />

interoperability artifact shall be. Hence, an interoperability artifact shall be defined as any artifact that is<br />

able to facilitate interoperability between information systems. This concept is defined in a wider sense as<br />

Fustar's Interoperability Meaning Diagrams indicates and fits also the widely used definition of<br />

interoperability as the capability of systems to exchange information (Institute of Electrical and Electronic<br />

Engineers 1990: 114). But within this paper only such interoperability artifacts will be considered that<br />

facilitate the interoperability of information systems managing or processing information about public<br />

services. In this sense, an artifact can be any application systems, software tools, system architecture<br />

modeling language, as well as analysis and design methods. These artifacts are characterized by a<br />

remarkable complexity by the underlying information technology, by a rapid technological change, and by<br />

a notional and conceptual diversity in naming and terms. The latter applies in particular for artifacts<br />

developed and used in the practice (Frank 2000: 39). If not stated otherwise and to simplifying the<br />

language the term interoperability artifacts will be used for interoperability artifacts used to exchange<br />

public service information within the further course of this paper.<br />

232


Veit Jahns<br />

3. Elaboration of the conceptual framework<br />

Usually evaluation is defined as determine the worth or the value of something using appropriate, explicated<br />

and justified criteria (Scriven 1991 in House 1993: 1). But speaking about worth and value, raises<br />

the question, for whom and to what end? And how should a framework used to perform the evaluation be<br />

designed and what are the reasons justifying the selected criteria? As it is the concern of this paper to<br />

evaluate artifacts, it should be drawn upon here on a proposal by Frank (2009: 41–43). As he is<br />

concerned in his paper with the evaluation of artifacts in the information systems research and which<br />

requirements should be fulfilled by an framework supporting the evaluation of artifacts, he proposes the<br />

following criteria (Frank 2009: 43):<br />

Goal-orientation: The objectives of the evaluation should be explicated and justified. In particular, to<br />

avoid or at least to reduce chance of an instrumentalistic fallacy (Albert 1960: 217), where the<br />

objectives—here the objective of an evaluation—is determined by the means—here the selected<br />

criteria (see e.g. Frank 2009: 37 or for a similar argumentation Streeten 1954).<br />

Seriousness: The scientific standard should not be sacrificed in favor of a high relevance to the<br />

practice. In particular, the criteria's operationalization should be made in a cautious manner.<br />

Justification: The proposed criteria should be justified in a comprehensible manner.<br />

Regarding the goal-orientation it is referred to the explanation and justification of this paper's objective<br />

given above. In the objective it is asked about interoperability artifacts, their scope of application as well<br />

as the underlying technologies and concepts. Regarding the seriousness of an evaluation of interoperability<br />

artifacts, it seems valuable to take a close look on the nature of interoperability artifacts<br />

regarding the exchange of public service information:<br />

Interoperability artifacts are created by humans in a development process. As a result of a development<br />

process the interoperability artifacts pass through a life cycle. They were created and<br />

applied, they mature and die, i.e., that they will not be used anymore. And because the development<br />

of interoperability artifacts can be considered as an eGovernment project, the development of an<br />

interoperability artifact has to deal also with the “ongoing changing pattern of relations or interactions<br />

between government organizations, business, and citizens” (Gupta and Jana 2003: 368).<br />

They describe more or less the public services. Thus, they are mostly some kind of models, codes,<br />

taxonomies, or in short: lingual artifacts.<br />

The subject—public services—of the interoperability artifacts considered in this paper, is characterized<br />

by a “fuzziness”, which makes it difficult to find the one interoperability artifact. Instead, a<br />

wide spectrum of manifestations of interoperability artifacts—as stated above—can be found.<br />

Due to the aspects given above many different perspectives and views have to be considered and/or<br />

incorporated into the interoperability artifact. But interoperability itself is a manifold phenomenon,<br />

which cannot be reduced to one single and isolated aspect (Dorloff et al. 2011).<br />

Considering these aspects, i.e., that interoperability artifacts are difficult to grasp and to quantify, that<br />

they are primarily lingual constructs as well as having Rowlands (2003: 3) characterization of qualitative<br />

research in mind that “qualitative research tends to work with text rather than numbers”, it seems appropriate<br />

to see interoperability artifacts as qualitative data and therefore to use an qualitative approach in<br />

this evaluation. But to give the qualitative evaluation a structure it will be recurred to an approach that has<br />

its origin more in the field of quantitative evaluation: the Goal/Question/Metric (GQM) paradigm (Basili<br />

1992). But here not the complete GQM paradigm will be applied, but rather its structure will be borrowed<br />

here.<br />

The GQM paradigm the procedure is that in a first step the objective or goal will be refined by deriving<br />

questions from the objective, and subsequently these questions will be made measurable by defining<br />

metrics. So, to refine the objective of this paper the questions will be formulated. To this particular end<br />

and as the issue of exchanging public service information between public authorities is related to interoperability<br />

of information systems a step back will be taken to look at this issue and to observed it from<br />

a greater distance. The step back taken is a step to the point of view by Dorloff et al. (2011). They<br />

examine in the paper the field of interoperability problems. To structure this field they refer to an idea<br />

once used by Wand and Weber (1995) and differentiate the phenomenon of interoperability from the<br />

social, communication, and the technological dimension (Dorloff et al. 2011: 5–6):<br />

233


Veit Jahns<br />

Social dimension: Here they refer to the aspects and characteristics of interoperability related to the<br />

way how humans experience and influence interoperability. As they consider interoperability from a<br />

wide perspective. Every aspect and characteristic of the social-economic reality is gathered here.<br />

Communication dimension: They consider in this dimension the linguistic aspects and characteristics<br />

of interoperability. Primarily, the aspects and characteristics of the artifacts created to achieve<br />

interoperable information systems.<br />

Physical dimension: In this dimension Dorloff et al. (2011) consider the technology used to achieve<br />

interoperability, e.g., the used protocols and middle-ware technologies.<br />

The categorization by Dorloff et al. (2011) based on the view, that interoperability is an object. As this<br />

object has different aspects and facets they approach this object by observing it from different perspectives<br />

each on highlighting some particular aspects and hiding others.<br />

Another and orthogonal approach to understand interoperability is, to understand interoperability as a<br />

process, i.e., that interoperability is established on a certain point of time, operated for a certain time span<br />

within some kind of collaboration until the collaboration ends and the interoperability will be dissolute.<br />

Based on this view Mallek et al. (2010: 441–446) developed an interoperability requirements model,<br />

which considers interoperability from the perspective of compatibility, inter-operation and reversibility<br />

requirements:<br />

As compatibility requirements they consider functionality, capability or characteristic of an information<br />

system that is related to its interoperability and is independent of time. Compatibility requirements<br />

need to be satisfied by organizations or its respective information systems to become interoperable.<br />

Inter-operation requirements are requirements related to the interactions of organization respective<br />

their information systems. How the interaction is performed and so on. So inter-operation<br />

requirements are dependent of time and primarily of a dynamic nature.<br />

Reversibility requirements refer to the end of a collaboration. In particular, this requirements aim to<br />

maintain the autonomy of an organization and reverse it to its original state after an collaboration has<br />

ended and no need to be interoperable with the collaboration organization is given. As the<br />

implementation of interoperability requires adaption and modification of the used information<br />

systems, the reversibility requirements aim to maintain the capability of an information system to<br />

return to its original state, e.g., by limiting the impact of the interoperability implementation on the<br />

information system itself.<br />

To combine the both views and to summarize the reflections above and to word the question it is referred<br />

to table 1. In this table the questions to be used to narrow down the objective of this paper are given.<br />

After the elaboration of the question, it is needed now to define the metrics. Instead of defining metrics for<br />

these questions the concept of morphology also called morphological box (Zwicky 1973: 94–100) is used<br />

here. This approach was originally a creative technique to find new and undiscovered instances of a<br />

given type of entities (solutions, types, artifact, etc.). It is looked for the essential characteristics of an<br />

entity type. The characteristics (rows) and its possible categories (columns) are arranged in a<br />

morphological matrix describing all possible objects of the entity type. By combining the applicable<br />

category—or categories—of each characteristic it is possible to describe the object considered object. At<br />

the first glance, it may be inappropriate here to use a creativity technique for an evaluation, but as the<br />

purpose of this evaluation is Also the concept of the morphological box has been successfully used before<br />

to examine similar research questions, e.g., by Dorloff et al. (2011) with respect to e-business<br />

interoperability. Furthermore and as the questions given above suggest, it is not the intention of this<br />

paper to perform a quantitative evaluation, but a rather to perform an evaluation in a qualitative way. The<br />

concept of morphology by Zwicky offers the opportunity to do this in a structured but also in a simple<br />

manner.<br />

By perusing the questions given above, the morphological box will be constituted by the following<br />

characteristics and categories:<br />

Question (1a) can be related to the communicating pattern between actors. These types of actors are<br />

can be classified as citizens (C), business (B), non-profit organizations (N) and public administration<br />

(A). As all these actors interact amongst themselves, only those interactions where the public<br />

234


Veit Jahns<br />

administration participates are considered here (Dorloff et al. 2011: 8; Sundgren 2005: 85; Wirtz<br />

2001: 34–39).<br />

Table 1: Evaluation questions<br />

Requirement<br />

Dimension<br />

Social Dimension (1a) Which type of actors<br />

exchange public service information?<br />

(1b) Which perspective of<br />

social actors are considered?<br />

(1c) Are legal rules taken into<br />

account?<br />

Communication<br />

Dimension<br />

Physical Dimension<br />

Compatibility Inter-operation Reversibility<br />

(4a) How is the concept of<br />

public service and its subconcepts<br />

defined in detail?<br />

(4b) What is the formalization<br />

degree of the artifact?<br />

(7a) What interoperability<br />

entities at the technological<br />

level are considered in the<br />

interoperability by the artifacts?<br />

(7b) Have and existing governmental<br />

interoperability<br />

frameworks been taken into<br />

account and if so, which one?<br />

(2) Does the artifact indicators,<br />

how this artifact<br />

improves the interoperability<br />

between two organizations<br />

respective information systems,<br />

e.g., by reducing costs,<br />

and how this improvement<br />

can be measures?<br />

(5a) What spectrum of the<br />

interaction process is<br />

covered by the interoperability<br />

artifact?<br />

(5b) Does it reflect the<br />

different types of information<br />

accrued during the interaction<br />

process?<br />

(8) Does the artifact indicators,<br />

how this artifact<br />

facilitate the implementation<br />

of interoperable information<br />

systems in public<br />

administrations?<br />

(3) What assumptions are<br />

made in the artifacts' creation<br />

process regarding the<br />

duration of the interaction?<br />

(6) Has the own terminology<br />

regarding public services to<br />

be adapted to the underlying<br />

terminology of interoperability<br />

artifact?<br />

(9) How tight is the integration<br />

of these artifacts intro the<br />

information systems?<br />

For question (1b) it referred to a distinction once made by Jahns et al. (2009: 359). They distinguished<br />

ontologies by the point of view from which the ontology was developed, i.e. the view of the<br />

public administration or the citizens’ respectively the business’ point of view.<br />

As the actions of the public administration is rule by law, which impedes the implementation of<br />

eGovernment services, the question (1c) aims at this fact, by asking, if legal rules are referable in the<br />

artifacts (Alpar and Olbrich 2005).<br />

The performance evaluation of interoperability artifact is a task of its own and not within the scope of<br />

this paper. But in the other side, many creators of interoperability artifacts promise that their<br />

interoperability artifact improves interoperability in one way or the other. So, it seems reasonable to<br />

ask, if indicators and possible metrics are provided with the interoperability artifacts. Like question<br />

(2), this is a binary question and two categories can be considered here: yes and no.<br />

From question (3) the characteristics of the interaction’s stability, i.e., if the interoperability artifacts<br />

tends more to support ad hoc interactions or long-term respectively permanent interactions.<br />

The type of definition of public service concept given in the respective approach can used to provide<br />

indications about the answer of question (4a). There can be found definitions based on public tasks<br />

as defined in public regulations like laws, or definition based on activities in public organizations<br />

(Jahns et al. 2009: 359).<br />

Regarding degree of formalization ask about in question (4b), there are three possible categories no<br />

formalization, i.e., using natural language, semi-formalization, .e.g., structured natural language or<br />

graphical models without formal semantics, and complete formalization with a well and formal defined<br />

semantic.<br />

For question (5a) and (5b) a characterization by Sundgren (2005: 90–92) of data processed in public<br />

information systems can be applied here: operative data, i.e., data that is necessary to process and<br />

235


Veit Jahns<br />

to complete as task within the provision of a public service, procedural data, i.e., descriptions or<br />

representations of rules related to the provision of public services, e.g., rules emanate from laws,<br />

directive or analytical data, i.e., statistics and other indicators regarding the state of the public services'<br />

provision, e.g., performance, meta-data, i.e., data describing the data itself, e.g., what format it<br />

is used to explicate the data, the source of the data and so on, process data, i.e., information<br />

gathered in the provision of the public services they inform about, e.g., archival data, i.e., data<br />

extracted from or generated during a public service provision to reflect or document the provision.<br />

Public authorities have their own special terminology regarding their field of activity. Also the<br />

interoperability artifacts require a fixed understanding of concepts they are dealing with. But is an<br />

adequate mapping between these terminologies possible or have the public authorities to adapt their<br />

terminology to the underlying terminology of the interoperability artifact? So, two answers to question<br />

(6) are possible here: yes and no.<br />

Question (7a) Here five categories of entities can be identified: the data types as the smallest and<br />

simplest entities dividing atomic data elements into different types and determining the allowed<br />

values of a data element, the vocabulary, i.e., the data elements defined on the basis of the data<br />

types, the documents as a set of logical related data, and the process as the sequences of<br />

exchanged documents (Dorloff et al. 2011: 10; Beckmann et al. 2004: 52–56).<br />

Question (7b) refers to sets of guidelines and regulations, e.g., the <strong>European</strong> Interoperability<br />

Framework (<strong>European</strong> Communities 2004) regarding the implementation of information systems in<br />

public authorities. Such regulations can be found on a national, international and global level.<br />

Similar to question (2), one can also ask with question (8), if indicators are provided allowing the<br />

determination, if and to what degree an interoperability artifacts facilitate the implementation of an<br />

information systems. As this is also a binary question the same two categories can be identified as in<br />

question (2).<br />

Refining question (9) two options are at hand here. Firstly, annotating existing public service<br />

information or export into a separate exchange format (Jahns et al. 2009: 359). Secondly, regarding<br />

the interoperability mechanism, i.e., achieving interoperability by standardization or by conversion.<br />

(Dorloff et al. 2011: 10; Wüstner et al. 2002: 55–57). Standardization means that allowing more<br />

individuality and therefore limits the impact on the information system and makes it easier to restore<br />

the original state. In contrast, standardization tends to have a stronger impact on information systems<br />

as its inner structure and processes may have to be changed to be compliant with the standard. A<br />

conversion approach on the other side allows preserving these inner structure and processes.<br />

These characteristics and their categories are summarized in table 2.<br />

Table 2: Morphological box for evaluation interoperability artifacts<br />

Characteristics <br />

Communicating<br />

Pattern<br />

Categories<br />

A2A A2B B2A A2C C2A A2N N2A<br />

View Citizens/Business Administration<br />

Considering<br />

Legal<br />

Frameworks<br />

Yes No<br />

Duration Ad hoc Long-term<br />

Public ServiceDefinition<br />

Degree of<br />

formalization<br />

Types of<br />

data<br />

Activity-based Task-based Not Explicated<br />

Natural Semi-formal Formal<br />

Operative Procedural Analytical Metadata Process data Archival data<br />

236


Characteristics <br />

Interoperabilityentities<br />

Considering<br />

Interoperability<br />

Frameworks<br />

Facilitation<br />

Indicators<br />

Interoperability<br />

mechanism<br />

4. Conclusion<br />

Veit Jahns<br />

Categories<br />

Data types Vocabulary Document Process<br />

National International Global<br />

Yes No<br />

Standardization Conversion<br />

Based on the premise, that there are many artifacts to support the interoperability of public administration<br />

with respect to the exchange of public service information, a conceptual framework for the evaluation of<br />

these artifacts was developed. The development of this framework was completely based on theoretical<br />

reflections on this issue. What is missing up to now is the application and use of this conceptual<br />

framework by the means of real interoperability artifacts? This will be the next step will. The framework<br />

will be applied to evaluate the existing artifacts created in research and implementation projects, e.g.,<br />

those mentioned at the beginning of this paper. But performing the evaluation will not only prove—or<br />

disprove—the framework itself. As the used perspectives to develop the framework highlight particular<br />

aspects of interoperability, and obscure others, performing the evaluation can reveal further perspectives<br />

and aspects of interoperability, which should be considered in an evaluation also. Furthermore, the result<br />

of the evaluation can be helpful to integrate these interoperability artifacts and reveal new research<br />

directions, as well as can be—a minor hope of the author—a contribution to the efforts (e.g. Charalabidis<br />

et al. 2010) to establish a science of interoperability or at least give interoperability a more profound scientific<br />

base.<br />

References<br />

Albert, H. (1960) “Wissenschaft und Politik. Zum Problem der Anwendbarkeit einer wertfreien Sozialwissenschaft”, in<br />

Topitsch, E. (ed.) Probleme der Wissenschaftstheorie. Festschrift für Victor Kraft, Springer, Wien.<br />

Alpar, P. and Olbrich, S. (2005) “Legal Requirements and Modelling of Processes in eGovernment”, Electronic<br />

Journal of eGovernment, Vol. 3, No. 3, pp. 107–116.<br />

Basili, V. R. (1992) Software Modeling and Measurement: The Goal/Question/Metric Paradigm, [online], Available:<br />

http://www.cs.umd.edu/~basili/publications/technical/T78.pdf [06 Nov 2010].<br />

Beckmann, H., van der Eijk, P., Hjulstad, H., Madsen, B. N., Ondracek, N. and Schmitz, V. (2004) CEN Workshop<br />

Agreement 15045. Multilingual Catalogue Strategies for eCommerce and eBusiness, [online], Available:<br />

ftp://ftp.cenorm.be/PUBLIC/CWAs/e-Europe/eCat/CWA15045-00-2004-Jul.pdf [21 Nov 2010].<br />

Charalabidis, Y., Gonçalves, R. J. and Popplewell, K. (2010) “Developing a Science Base for Enterprise<br />

Interoperability”, in Popplewell, K., Harding, J., Poler, R. and Chalmeta, R. (ed.) Enterprise Interoperability IV.<br />

Making the Internet of the Future for the Future of Enterprise, Springer, London.<br />

Dorloff, F.-D., Jahns, V. and Schmitz, V. (2011) “E-Business Interoperability: A Systematization Attempt Based on<br />

the Morphology Concept”, [forthcoming], in Kajan, E. (ed.) Electronic Business Interoperability: Concepts,<br />

Opportunities and Challenges, IGI Global, Hershey.<br />

<strong>European</strong> Communities (2004) <strong>European</strong> Interoperability Framework for Pan-european EGovernment Services,<br />

[online], Available: http://ec.europa.eu/idabc/servlets/Docd552.pdf?id=19529, [21 Nov 2010].<br />

<strong>European</strong> Union (2009) Ministerial Declaration on eGovernment, [online], Available: http://www.se2009.eu/polopoly_fs/1.24306!menu/standard/file/Ministerial%20Declaration%20on%20eGovernment.pdf<br />

[15 Oct 2010].<br />

Frank, U. (2000) “Die Evaluation von Artefakten: Eine zentrale Herausforderung der Wirtschaftsinformatik”,<br />

Häntschel, I. and Heinrich, L. (ed.) Evaluation und Evaluationsforschung in der Wirtschaftsinformatik,<br />

Oldenbourg, München.<br />

Fugge, F. (2003) “Administrative Traditions in Western Europe”, in Peters, B. G. and Pierre, J. (ed.) The SAGE<br />

Handbook of Public Administration, Sage, London.<br />

Fustar, S. (2009) “A Generic Approach for Compliance Assessment of Interoperability Artifacts”, Grid Interop 2009<br />

Proceedings. The Road to an Interoperable Grid, Denver, CO.<br />

Garfield, E. (1977) “Why don't we have science reviews?”, in Essays of an Information Scientist, Vol. 2 (1974-1976),<br />

ISI Press, Philadelphia.<br />

237


Veit Jahns<br />

Gupta, M. P. and Jana, D. (2003) “EGovernment Evaluation: A Framework and Case Study”, Government<br />

Information Quarterly, Vol. 20, No. 4, pp. 365–387.<br />

House, E. R. (1993) Professional Evaluation. Social Impact and Political Consequences, Sage, Newbury Park.<br />

Institute of Electrical and Electronic Engineers (1990) IEEE Standard Computer Dictionary. A Compilation of IEEE<br />

Standard Glossaries, New York.<br />

Jäckering, W. (1994) “Verwaltung und öffentliche Aufgaben”, in Mattern, K.-H. (ed.) Allgemeine Verwaltungslehre, 4 th<br />

edition, Walhalla und Praetoria, Berlin.<br />

Jahns, V., Dorloff, F.-D. and Schmitz, V. (2009) “Harmonising and Exchanging Public Service Information Between<br />

Public Service Centres. Requirements, Constraints and Outline of a Solution”, in Hahamis, P. (ed.) Proceedings<br />

of the 9 th <strong>European</strong> <strong>Conference</strong> on eGovernment, London, pp. 377–384.<br />

Lynn, L. E. (2006) Public Administration. Old and New, Routledge, New York.<br />

Mallek, S., Daclin, N. and Chapurlat, V. (2010) “Towards a Conceptualisation of Interoperability Requirements”, in<br />

Popplewell, K., Harding, J., Poler and R. Chalmeta, R. (ed.) Enterprise Interoperability IV. Making the Internet of<br />

the Future for the Future of Enterprise, Springer, London.<br />

Rowlands, B. (2003) “Employing Interpretive Research to Build Theory of Information Systems Practice”, in<br />

Australasian Journal of Information Systems, Vol. 10, No. 2, pp. 3–22.<br />

Scriven, M. S. (1991) Evaluation Thesaurus, 4 th edition, Sage, Newbury Park.<br />

Streeten, P. (1954) “Programs and Prognoses”, The Quarterly Journal of Economics, Vol. 68, No. 3, pp. 355–376.<br />

Sundgren, B. (2005) “What is a Public Information System?”, in International Journal of Public Information Systems,<br />

Vol. 1, No. 1, pp. 81–99.<br />

Wand, Y. and Weber, R. (1995) “On the Deep Structure of Information Systems”, Information Systems Journal, Vol.<br />

5, No. 3, pp. 203–223.<br />

Wirtz, B. W. (2001) Electronic Business, 2 nd edition, Gabler, Wiesbaden.<br />

Zwicky, F. (1973) Entdecken, Erfinden, Forschen im morphologischen Weltbild, Droemer Knaur, München.<br />

Wüstner, E., Hotzel, T. and Buxmann, P. (2002) „Converting Business Documents: A Classification of Problems and<br />

Solutions Using XML/XSLT“, in Proceedings of the 4 th IEEE International Workshop on Advanced Issues of E-<br />

Commerce and Web-Based Information Systems, Newport Beach, CA, pp. 54–61.<br />

238


Free and Open Source Software for Public Sector Enterprise<br />

Applications in Sri Lanka<br />

Srimal Jayawardena 1 and Gihan Dias 2<br />

1 Australian National University, Canberra, Australia<br />

2 University of Moratuwa, Sri Lanka<br />

srimal.jayawardena@anu.edu.au<br />

gihan@uom.lk<br />

Abstract: This paper describes a study done on the use of Free and Open Source Software (FOSS) for enterprise<br />

applications of the public sector in Sri Lanka. The study investigates factors that affect the selection of software for<br />

public sector information systems with, an emphasis on how Free and Open Source Software (FOSS) is used and<br />

could be used in such projects. The objectives of the study are as follows. The study aims to identify factors that affect<br />

the adoption of Software in Sri Lankan government sector projects. In addition to this it identifies features that<br />

are considered as important in such large Information Systems, and how they relate to Open Source Software adoption<br />

in the context of the public sector in Sri Lanka. The findings are analyzed and suitable recommendations are<br />

presented for better selection of software in the public sector. The work is important as it identifies and highlights<br />

factors that affect the choice of software in the public sector. This is important for several categories of people. It is of<br />

importance to the strategic management and policy makers to know what drives the information system procurement<br />

decisions in order to make more relevant policies and guidelines that are congruent with the needs in government<br />

sector departments. It is useful when advocating new technologies and information systems. This is especially true<br />

when advocating the use of Free and Open Source software for the use in the public sector in a wholesome and sustainable<br />

manner. The research is also important to software vendors and solution providers to the public sector in<br />

identifying what factors need to be taken in to account when bidding for public sector IS projects. The research is of a<br />

qualitative nature. It consists of multiple case studies of selected government sector departments and projects in Sri<br />

Lanka. The study investigates information systems developed by internal staff, developed by external consultants,<br />

procured systems, system implementation via private public partnerships (PPP) and projects guided by the Information<br />

and Communication Technology Agency (ICTA) of Sri Lanka. Data was gathered through interviews of staff at<br />

different levels of selected government sector Information Systems projects. The data was comparatively analyzed<br />

on a case by case basis to identify common patterns and trends among the investigated organizations and projects.<br />

The identified factors affecting the choice of software include the cost, technical specifications, bidder’s expertise in<br />

selected technology and user’s expertise in selected technology amongst other factors described in the full paper.<br />

The study identifies patterns between the choice of software - FOSS or non FOSS, and these identified factors.<br />

Based on these, recommendations are given to adopt and benefit from the use of FOSS in public sector enterprise<br />

level software projects.<br />

Keywords: enterprise software, government, free and open source software, software adoption<br />

1. Introduction<br />

We introduce the work to the reader by describing the research problem and objectives of the research.<br />

The research methodology, scope and nature of the study are briefly highlighted before being discussed<br />

in detail in the sections which follow.<br />

1.1 The research problem<br />

Government sector Information Systems projects guided by the ICTA (ICTA 2008) are a part of the e-<br />

Development initiative of Sri Lanka (e-Sri Lanka), which is funded primarily by the World Bank. The World<br />

Bank publication by Hanna (2006) describing the experiences of envisioning and designing of the Sri<br />

Lankan e-Development program highlights that it can benefit from the use of FOSS. Apart from this, the<br />

literature surveyed has highlighted how organizations could benefit from the use of FOSS.<br />

However, to realize the benefits of FOSS, decision makers should perceive it as a viable option. Hence it<br />

is important to identify what affects the choice of software at a strategic level. This leads us to the problem<br />

statement of this research.<br />

“What factors influence the choice of software for government sector enterprise applications in Sri<br />

Lanka?”<br />

Complexity. According to (Lewin 2000), a complex system is one whose properties are not fully explained<br />

by an understanding of its parts.<br />

This problem statement is further elaborated by the research objectives explained below.<br />

239


1.2 Research objectives<br />

Srimal Jayawardena and Gihan Dias<br />

Based on the above problem statement, this research consists of the following objectives.<br />

1. To identify factors that affects the adoption of Software in Sri Lankan government sector projects.<br />

2. To identify what features are considered as important in large Information Systems, and how they<br />

relate to Open Source Software.<br />

3. To develop suitable recommendations for the better selection of software in the Sri Lankan government<br />

sector.<br />

1.3 Significance of the study<br />

The study identifies the factors that affect the adoption of software for enterprise applications. Moreover,<br />

it focuses on those factors that are peculiar to the government sector in Sri Lanka. It also identifies what<br />

aspects of ISs are taken in to account when different government sector institutions evaluate an IS. The<br />

study will help formulate polices that will aid government sector organizations to leverage on the benefits<br />

of FOSS as explained in section 2.<br />

1.4 Research methodology<br />

The research methodology involved a rigorous study of selected government organizations as out lined in<br />

the scope of the study. The research was designed based on the findings of the literature study and the<br />

research objectives. Data collection was carried out using personal interviews based on this research<br />

design. The obtained data has been analyzed in detail in section 4. The research methodology is described<br />

in depth in section 3.<br />

1.5 Scope of the study<br />

The scope of the research is limited to government sector ISs. Government projects coordinated through<br />

the Information and Communication Agency (ICTA) of Sri Lanka and other selected government organizations<br />

have been taken as case studies (Appendix A & B).<br />

1.6 Data collection<br />

The organizations covered included, those with ICTA guided projects as well as organizations without an<br />

ICTA involvement for comparison purposes (Appendix A). The data was collected by interviewing staff at<br />

different levels in the selected organizations (Appendix A). This included personnel at a strategic level<br />

and heads of the IT/IS departments. In the case of ICTA guided projects the interviews consisted of the<br />

ICTA project managers and relevant external stakeholders. The information systems (IS) projects covered<br />

in the research are indicated in Appendix B. Interviewed personnel and sample interview questions<br />

are included in Appendices C and D.<br />

1.7 Data analysis<br />

The data was comparatively analyzed on a case by case basis to identify common patterns and trends<br />

among the investigated organizations and projects.<br />

1.8 Nature and form of results<br />

The results of the research are qualitative in nature. It is a comprehensive study of the factors that affect<br />

the adoption of FOSS in the Sri Lankan government sector organizations for enterprise systems. The<br />

results also include suitable recommendations for capitalizing on the advantages of FOSS in the Sri<br />

Lankan government sector.<br />

1.9 Structure of this document<br />

The remainder of this paper is organized as follows. A background to the problem is presented which is<br />

followed by an explanation of the research methodology employed. This followed by an analysis of the<br />

investigation outcomes which are followed by conclusions and recommendations.<br />

240


2. Background<br />

Srimal Jayawardena and Gihan Dias<br />

We present a background to the work by consolidating literature surveyed on Free and Open Source<br />

Software (FOSS), on the use of FOSS in the government sector and on FOSS adoption models. FOSS<br />

is claimed to have a unique blend of benefits that have aroused the interest of organizations across the<br />

world. Some of these benefits have been reported to be as follows.<br />

A Total Cost of Ownership (TCO) study conducted by the Robert Francis Group (Orzech 2002) reveals<br />

that the use of FOSS computer operating systems like GNU/Linux can lower TCO up to 40%.<br />

Further to this, Gartner has reported a cost reduction of 15% (Maguire 2003) when using FOSS Operating<br />

Systems as opposed to Windows XP, which is a proprietary Operating System. A study done<br />

by Merill Lynch (Lemos 2003), a major financial management company, has revealed that TCO reductions<br />

by the use of FOSS are not only due to software licensing costs but from personnel and<br />

hardware costs.<br />

A recent e-Primer of the United Nations Development Program (Wong 2004) mentions enhanced<br />

security and vendor independence as other benefits of the use of FOSS in government applications.<br />

2.1 Free and Open Source Software (FOSS)<br />

According to the Wikipedia (2005), Free and Open Source Software are software that is licensed such<br />

that its human readable source code is freely available for implementers to study, change and improve its<br />

design.<br />

Research (Reddy and Evans, 2002) reveals that more than 48% of the government institutions in<br />

Chile use FOSS with at least another 11% planning to do so. Out of this 93% of the institutions use<br />

FOSS for server side applications. According to the findings of Chan (2007), several developing governments<br />

like Peru have attempted to adopt FOSS at a national level.<br />

A recent online survey done by the CIO magazine (Schindler 2008) during the period April 28, 2008<br />

to May 2, 2008 shows that more than 50% of the participants are using FOSS. Another 10% of the<br />

participants are planning to use FOSS during the next 12 months. Therefore there appears to be a<br />

growing interest in the use of FOSS across the world.<br />

10.00%<br />

53.00%<br />

Figure 1: Survey conducted by the CIO magazine<br />

37.00%<br />

241<br />

Currently<br />

using<br />

Planning to<br />

use next 12<br />

months<br />

No plans to<br />

use open<br />

source<br />

applications


2.2 Success in Enterprise Systems<br />

Srimal Jayawardena and Gihan Dias<br />

Several models have been developed to asses the success of Enterprise Systems. Out of these the<br />

popular D & M model (DeLone and McLean 2003) considers the success of the system based in the information<br />

quality, system quality and the service quality. The model strives to analyze the associations<br />

between these dimensions and the intention to use, the user satisfaction and net benefits. A pictorial representation<br />

of the updated D & M model is shown in the figure 2.<br />

Figure 2: The updated D & M model<br />

The Enterprise System Success Model (Gable et al 2003) presents a temporal analysis of the success<br />

based on what is to date and what happens in the future. The success to date is said to be influenced by<br />

the organizational impact and the individual impact. The future success is said to be influenced by the<br />

system quality and the information quality. A pictorial representation of the Enterprise Success Model is<br />

given in figure 3.<br />

2.3 Use of FOSS in the government sector<br />

Government sector software applications require long term continuity and a lower total cost of ownership<br />

(TCO) while maintaining a high degree of quality and integrity. Mitch Stoltz (1999) mentions that governments<br />

can use open source as a vehicle for promoting economic development.<br />

Some of the benefits of the use of FOSS in the government sector have been identified by Reddy and<br />

Evans (2004) as follows.<br />

2.3.1 Cost savings<br />

Significant cost savings can be made due to the lack of expensive licensing fees in FOSS (Reddy and<br />

Evans 2004). However this factor alone does not justify the use of FOSS.<br />

2.3.2 Security, stability and privacy<br />

As the source code is open for inspection, it is claimed that FOSS is more secure and stable. According<br />

to Reddy and Evans (2004), the German government was concerned that some versions of the proprietary<br />

Microsoft Windows operating system contained back doors designed to grant the US National Security<br />

Agency, access to user data. Based on this rationale, the government of Germany has adopted a<br />

strong FOSS policy.<br />

2.3.3 Independence<br />

Proprietary software solutions are often based on one or few software companies. Hence a government<br />

implementing such proprietary solutions will have to depend on those few companies for support and<br />

business continuity. Due to the open nature of FOSS, dependence on software companies is less.<br />

242


Figure 3: The enterprise success model<br />

Srimal Jayawardena and Gihan Dias<br />

2.3.4 Helping domestic industries<br />

The open nature of FOSS will help the domestic software industry to actively contribute in developing<br />

applications. This is important for the long term sustainability of government projects and in terms of national<br />

interests.<br />

2.3.5 Innovation<br />

The open nature of FOSS helps innovation. This results in more advanced and efficient software which<br />

are released faster than proprietary software which have to adhere to conventional release cycles.<br />

Figure 4: Kwan and West (2005) FOSS adoption model<br />

2.4 Open Source adoption models<br />

Kwan and West (2005) of the Silicon Valley Open Source Research Project, San José State University,<br />

have presented the following Conceptual Model for Enterprise Adoption of Open Source Software (Figure<br />

4).<br />

The model assumes a three phase process:<br />

2.4.1 Policies<br />

The firm establishes a series of policies (formal or informal) that guide its procurement decisions.<br />

243


2.4.2 Choice set<br />

Srimal Jayawardena and Gihan Dias<br />

From the range of available options, at the first decision point the firm identifies those acceptable alternatives<br />

that meet minimum requirements.<br />

2.4.3 Selection<br />

From the range of acceptable alternatives, at the second decision point, the firm makes its final selection<br />

decision.<br />

The following factors are said to influence the adoption decisions.<br />

Features<br />

These are the attributes commonly used to measure what is new or valuable about a given technology,<br />

and used by vendors to differentiate their products. Because anything could be tautologically classified as<br />

a “feature,” we define feature as the residual of the two remaining categories<br />

Risk<br />

Risk for a system is often thought of in terms of reliability — the comparative scarcity of crashes, failures,<br />

or data loss. However, risk-lowering measures might also include efforts to mitigate the effect of inevitable<br />

if rare failures (such as redundant data or 7/24 support). Firms may also consider the risks of their<br />

investments over time, such as the risk that the technology may be orphaned by the vendor (either deliberately<br />

or with the vendor’s bankruptcy), or that for some other reason the investment is rendered obsolete.<br />

Cost<br />

These may include both the initial purchase price, and the ongoing usage costs such as support contracts<br />

and upgrade fees. An attempt to calculate total cost of ownership would also include personnel<br />

cost, and (for large data centers) related equipment costs such as power, air conditioning, security, etc.<br />

The following factors have been identified by Dedrik and West (2003) as open source adoption factors in<br />

their grounded theory based research.<br />

Willingness to take risks on a new, unproven technology.<br />

Need for organizational slack to evaluate the new technology and to self-support unsponsored technologies.<br />

Tendency of open source software to be inexpensive if not free and the inherent trialability of “free”<br />

software distributed on the Internet.<br />

Availability of external sources of support and expertise.<br />

Having set the background to the research with the aid of relevant literature, we move on to describe the<br />

research methodology employed in the section which follows.<br />

3. Research methodology<br />

The study intends to investigate a selected set of government IS in considerable depth. Therefore a case<br />

study based approach was adopted. As suggested by Yin (2003) the overall evidence from a multiple<br />

case study could be considered to be more compelling and robust, when it can be afforded to. Therefore,<br />

selected enterprise IS of selected government organizations were taken as case studies. Therefore the<br />

study is mostly qualitative in nature. The data collection was done through in-depth interviews. Staff at<br />

different levels of the organization has been interviewed as appropriate in order to obtain a more holistic<br />

perspective. The selected government organizations (Appendix A), the IS projects (Appendix B), personnel<br />

interviewed (Appendix C) and the interview outline with sample questions (Appendix D) are included<br />

at the end of the paper.<br />

244


3.1 Scope of the research<br />

Srimal Jayawardena and Gihan Dias<br />

The research focuses on IS pertaining to a selected set of government organizations. The sample set<br />

consists of ICTA guided IS projects as well as non-ICTA IS projects in order to facilitate comparative<br />

analysis. The list of organizations and the IS projects included in the study are indicated in Appendix A<br />

and B.<br />

3.2 Interview design<br />

In depth interviews were carried out based on a broad outline. The structure of the interviews was designed<br />

to bring out the factors that affect the choice of software in the organizations concerned. In particular,<br />

the selection or the absence of FOSS has been investigated. The interviews investigate the<br />

source of the requirements specifications along with the nature of the technologies and factors that influenced<br />

the IS solution. The IS solution was examined in terms of its modules which included different<br />

combinations of FOSS and proprietary technologies. This ranged from totally FOSS or proprietary systems<br />

to IS with a combination of both. Issues and room for improvement in the selected IS have also<br />

been investigated. The interview outline covers the following aspects.<br />

3.2.1 A general investigation<br />

The main projects, the project objectives and components were identified in this section.<br />

3.2.2 Technology and components<br />

The software, hardware and middle ware components used for each project were identified. The use of<br />

proprietary and FOSS components were investigated and identified.<br />

3.2.3 Specification development and selection<br />

The nature of the specifications development for each IS was identified. This section investigates if specifications<br />

were developed internally, by an external consultant and in the case of ICTA guided projects,<br />

the involvement of a SAGE.<br />

3.2.4 Factors affecting the selection<br />

This section investigates the key factors affecting the selection of the IS system concerned. In the case of<br />

a procured solution the factors affecting the selection process were investigated. It was also investigated<br />

if the interviewees had any recommendations to improve the current selection process. If the IS was developed<br />

by internal staff, the reasons behind this decision was investigated.<br />

3.2.5 Issues, improvements and FOSS<br />

For ISs that have already been fully implemented, the issues and room for improvement in those systems<br />

were investigated. Also the management inclination towards using FOSS components as a viable option<br />

to resolve those issues and to make improvements was investigated. The perceived areas where FOSS<br />

could be used most advantageously have been investigated.<br />

The data collected using the described research methodology is analyzed in the next section.<br />

4. Analysis<br />

The following factors have been broadly identified to affect the selection of IS's, as a result of the in-depth<br />

interviews carried out during the study.<br />

Cost<br />

Technical Specifications<br />

Usability<br />

Bidders Expertise in selected Technology<br />

Users Expertise in selected Technology<br />

Interoperability and Open Standards<br />

245


Trilingual Support<br />

Support and Maintenance<br />

Security<br />

Srimal Jayawardena and Gihan Dias<br />

It has been attempted to analyze the factors based on how the requirements specifications were developed.<br />

The following categories have been identified during the study as possibilities of developing the<br />

requirements specifications.<br />

Developed by internal Staff<br />

Developed by External Consultant<br />

ICTA Guided<br />

Private Public Partnerships (PPP)<br />

4.1 Analysis of the factors<br />

The identified factors can be analyzed from several perspectives. An analysis of the 4 most commonly<br />

mentioned factors is shown in Figure 5. Based on this, organizations that considered ‘Bidders/Developers<br />

expertise in the selected technology’ as an important factor had a significantly lower percentage of FOSS<br />

systems as opposed to the others. Therefore, we see that this factor is a strong deterrent for using FOSS<br />

in the organizations investigated.<br />

Figure 5: Factor summary showing the use of FOSS and proprietary IS<br />

4.1.1 Cost<br />

Almost all investigated organizations have indicated the cost of the IS as a decisive factor. Almost half<br />

(43%) of investigated projects where the cost was considered as an important factor were for systems<br />

that were developed in-house. Next in line were ICTA guided projects which composed of 11% of the<br />

population. A few interesting points have been noted when analyzing the cost as a factor.<br />

Total Cost vs. IS Cost<br />

246


Srimal Jayawardena and Gihan Dias<br />

Certain projects, where the cost of the IS is small compared to the capital cost of the entire project have<br />

placed a lower importance on the cost of the IS. For example, in the CEB power systems the total cost of<br />

a power station is very large when compared to its IS component.<br />

80-20 Rule<br />

In most ICTA guided projects where the cost is a decisive factor, the cost of the projects has been given<br />

prominence in the requirements specification by using the 80-20 rule. This is where the final selection is<br />

done by allowing an 80% weight on the cost of the bid while a weight of 20% is given on technical specifications.<br />

This rule has been used to indicate to bidders that cost is an important factor in the selection<br />

process. In such situations, certain vendors have considered FOSS components in the proposed solutions.<br />

4.1.2 Technical compliance<br />

Technical compliance has been reflected in most projects by indicating a minimum level of compliance.<br />

Other factors have been considered subject to this minimum level of technical compliance. Certain projects,<br />

however, have specific technical requirements.<br />

Compliance with existing technology<br />

Certain projects impose technical constraints as it is required to migrate data that is already stored using<br />

a particular technology. For example the Department of Education has had its data in an IBM mainframe<br />

since 1974. This system was migrated to the IBM AS/400 platform in 1998 as the migration of the past<br />

data was easier within the selected platform.<br />

4.1.3 Usability<br />

Usability has been identified as a decisive factor in projects developed internally and in procured systems.<br />

However, usability was not explicitly highlighted as a decisive factor in the other categories of<br />

specification development.<br />

4.1.4 Implementers and specifications developers influence<br />

We see in figure 6 that different methods of specifications development and implementation have resulted<br />

in varying levels of FOSS and proprietary software being used in the resulting IS. The largest percentage<br />

of FOSS usage is reported from ICTA guided government enterprise IS projects. This reflects an<br />

increasing FOSS awareness within the ICTA. Other methods of implementations have lower levels of<br />

FOSS – lower than 50%. It is interesting to note that the PPP category consists entirely of proprietary<br />

software. Salient aspects of each of these categories are discussed next.<br />

PPP Projects<br />

In the Government Information Center (GIC) project, the private partner organization maintained that the<br />

single most reason for their choice for a particular set of proprietary technologies was because their expertise<br />

lay in those technologies.<br />

ICTA Guided Projects<br />

The Lanka Government Network (LGN) project is facilitated via a Korean government loan. The loan<br />

conditions dictate that the bidders of the project are required to be Korean. The expertise of these bidders<br />

has influenced the ultimate choice of technology for this project, although the ICTA has not limited the<br />

requirements to technology in particular. Also the degree on inter dependency between components of<br />

this project limit the freedom of choice given that a particular component has been implemented by a certain<br />

vendor using a proprietary solution.<br />

External Consultant<br />

The same notion is reflected under this category as well. It was found that a particular MIS at the Tertiary<br />

and Vocational Education Commission has been developed in a certain technology as the consultant<br />

247


Srimal Jayawardena and Gihan Dias<br />

concerned was familiar with it. The system has had problems in scaling up to new requirements that have<br />

evolved and a new MIS is being implemented at a larger scale.<br />

Figure 6: Specifications developer / Implementer effects on the IS<br />

4.1.5 Users expertise in selected technology<br />

The expertise of existing users in a given technology has been found to be an important factor in certain<br />

ICTA guided projects. This is particularly so in the LGN project which also relies on the expertise of the<br />

existing staff in government organizations to maintain the infrastructure installed at those organizations.<br />

4.1.6 Interoperability and open standards<br />

Compliance with open standards and interoperability has been identified as important factors in certain<br />

ICTA guided projects.<br />

4.1.7 Trilingual support<br />

Trilingual support has been highlighted as an important factor in most recent government sector projects<br />

that involve the general public as users. This was especially noted in the design of the E-Local Government<br />

project.<br />

4.1.8 Support and maintenance<br />

Support and maintainability of the IS projects have been especially highlighted as an important factor in<br />

the internally developed category. In the TVEC MIS project for example, the consultant who implemented<br />

the initial MIS had been unable to provided support and maintenance of the system after a certain point.<br />

Hence the management had decided that the future MIS should be implemented with internal expertise<br />

using an open technology, so that the system can be maintained by the internal staff.<br />

4.1.9 Security<br />

Security of the IS has been explicitly identified as an important factor in the Examination Information System<br />

of the Department of Education. In fact, security has been a strong factor for the choice of the particular<br />

technologies used to implement the system – namely IBM AS/400 based technologies.<br />

Having analyzed the data as explained above, we draw the following conclusions.<br />

248


5. Conclusions<br />

Srimal Jayawardena and Gihan Dias<br />

We conclude from the analysis that although a considerable number of government organizations perceive<br />

the benefits of using FOSS, the following factors prevent FOSS adoption.<br />

5.1 Lack of FOSS expertise in companies that bid for government projects<br />

A significant reason for not using FOSS based IS in the government sector was the fact that companies<br />

bidding for such projects did not propose such solutions. It is only recently that companies bidding for<br />

government projects have started to integrate FOSS components in the proposed solutions, to lower cost<br />

and leverage on the other benefits of FOSS.<br />

5.2 Lack of internal expertise in FOSS<br />

Certain organizations that were investigated noted that they do not have the expertise to implement and<br />

maintain FOSS IS internally. Therefore there is reluctance towards implementing FOSS solutions.<br />

5.3 Lack of FOSS support companies<br />

In several organizations, FOSS solutions have been implemented by internal software developers. A<br />

cause for concern was the lack of FOSS support companies to obtain support services. Often the FOSS<br />

expertise was from a few key developers in the staff who have had an exposure to FOSS.<br />

5.4 Use of unlicensed proprietary software<br />

The use of unlicensed proprietary software makes the cost unrealistically lower than when using legal<br />

licensed versions. This is a disincentive to consider FOSS alternatives which would have been otherwise<br />

cheaper.<br />

6. Recommendations<br />

Based on these conclusions, we make the following recommendations, in order to leverage on the benefits<br />

of FOSS in government sector enterprise software applications.<br />

6.1 Nurture a FOSS ecosystem<br />

FOSS service provider companies should be nurtured and encourage to support government IS development.<br />

This should encourage places where FOSS expertise exists, for example Universities, to provide<br />

FOSS services to government IS projects. Since, government IS will have long term benefits by leveraging<br />

on FOSS components, nurturing a FOSS ecosystem with companies that provide FOSS solutions will<br />

be a win-win situation for the government as well as for those companies.<br />

6.2 Developing internal expertise<br />

Internal expertise within the government organizations should be developed. IT personnel within government<br />

organizations should be trained to be FOSS savvy. Technical expertise within the organizations<br />

should be capable of maintaining FOSS IS. The following aspects could be considered in achieving this.<br />

6.2.1 Knowledge sharing<br />

A knowledge sharing mechanism among government (and other) organizations should be encouraged.<br />

This is to facilitate support and maintenance experience of FOSS solutions which would otherwise remain<br />

in silos.<br />

6.2.2 Government IT service<br />

With the increasing use of IT/IS within the government sector, a government IT service should be developed<br />

that could cater to the IS needs of government organizations. The personnel in this service should<br />

be trained to be FOSS savvy. There should be close liaison between this service and the private FOSS<br />

solution companies.<br />

249


6.3 Licensed software<br />

Srimal Jayawardena and Gihan Dias<br />

Licensed software should be insisted on when proprietary software components are used in government<br />

enterprise software applications. This will safeguard organizations from litigation when the legal framework<br />

develops and is capable of prosecuting the use of illegal software. Insisting on licensed software will<br />

make the costs of proprietary software components reflect realistically in the procurement process.<br />

Therefore, the monetary benefits of any comparable FOSS components will be more obvious.<br />

6.4 Policies in general<br />

Government organizations should encourage FOSS friendly policies in IS procurement policies. FOSS<br />

experience of companies is relatively young as FOSS itself is comparatively new. Government tender<br />

policies should accommodate this fact, if FOSS based solutions are to be encouraged. Also, open standards<br />

and non restrictive policies should be encouraged.<br />

List of Acronyms<br />

CIO – Chief Innovation Officer<br />

CRM – Customer Relationship Management<br />

FOSS – Free and Open Source Software<br />

HRM – Human Resource Management<br />

ICTA – The ICT Agency of Sri Lanka<br />

IS – Information System<br />

LGN – Lanka Government Network<br />

MIS – Management Information System<br />

PPP – Public Private Partnership<br />

SAGE – Software Architecture Group of Experts<br />

Appendix A<br />

Organizations investigated in the research are as follows.<br />

Table 1: Organizations investigated<br />

Organization Code<br />

Ceylon Electricity Board CEB<br />

Department of Agriculture DeptAgri<br />

Department of Education DeptEdu<br />

Department of Local Governments LocalGov<br />

ICT Agency of Sri Lanka ICTA<br />

Ministry of Public Administration PubAdmin<br />

Sri Lanka Customs Customs<br />

State Engineering Corporation SEC<br />

Tertiary and Vocational Education Commission TVEC<br />

Appendix B<br />

List of IS projects investigated are as follows. Please refer Appendix A for organization codes.<br />

250


Table 2: Investigated IS projects<br />

Appendix C<br />

Srimal Jayawardena and Gihan Dias<br />

Project Organization<br />

Accounting and Inventory System CEB<br />

Bill of Quantities System SEC<br />

Call Center DeptAgri<br />

Consumer Billing System CEB<br />

CRM System GIC/ICTA<br />

E-Learning System TVEC<br />

E-Learning System DeptAgri<br />

EHRM System ICTA/ PubAdmin<br />

eLocal Government Project ICTA/LocalGov<br />

ePensions Project ICTA/PensionsDept<br />

eSamurdhi Project ICTA/SamurdhiAuthority<br />

Examination Processing System DeptEdu<br />

Goods Processing System Customs<br />

Grama Niladari HR System PubAdmin<br />

Inventory System SEC<br />

LGN Phase 1 ICTA<br />

Mail Server SEC<br />

Management Information System TVEC<br />

Motor Vehicle Records System Customs<br />

Payroll System SEC<br />

Power System Control Systems CEB<br />

Timex BPO (Pvt) Ltd.(Public Private Partnership) PPP<br />

Warehouse Monitoring System Customs<br />

Web Information System DeptAgri<br />

Web/Email System PubAdmin<br />

People interviewed in different organizations. Please refer Appendix A for organization codes.<br />

Table 3: Interviewed personnel<br />

Name Designation Organization<br />

Dr. T.A Piyasiri Director General TVEC<br />

Mr. Ajantha Dias Director IT Customs<br />

Mr. Bandula Perera Chief Innovation Officer LocalGov<br />

Mr. Chinthaka Ranasinghe Project Manager ICTA<br />

Mr. Dammika Karangoda Chief Innovation Officer DeptEdu<br />

Mr. Janaka Jayalath Director Information Systems TVEC<br />

Mr. K.S.P Jayawardena Additional General Manager CEB<br />

Mr. Kumudu Munasinghe Head of Information Management SEC<br />

Mr. Lalith Waduge Systems Engineer PubAdmin<br />

Mr. M.G Tillakaratne Deputy General Manager CEB<br />

Mr. Moin Habeeb Manager PPP<br />

Mr. Neil Gunadasa Director of Education DeptEdu<br />

Mr. W.A.G Sissirakuimara Subject Matter Specialist<br />

Program Director- Re-engineering Govern-<br />

LocalGov<br />

Mr. Wasantha Deshapriya<br />

ment ICTA<br />

251


Appendix D<br />

Srimal Jayawardena and Gihan Dias<br />

The interviews were based on the following sample questions.<br />

1. What are the Information Systems (IS) projects in your organization and what are the components<br />

of those projects?<br />

2. What technology, hardware, software or middleware components were used in these IS? Were<br />

any FOSS components used?<br />

3. How were the specifications of the project developed?<br />

4. Who are the key people involved in the selection decision?<br />

5. Regarding the procurement process, a) What factors influence the selection decision in your procurement<br />

process? b) How do you rank these factors based on their importance for the existing procurement<br />

process? c) Do you have any recommendations to improve the selection process?<br />

6. Regarding issues and room for improvement, a) If these systems are up and running, are there<br />

any problems? b) In the said systems what areas need improvement? c) Which parts of it are you<br />

happy with and which parts are you not satisfied? d) For the unsatisfied areas, do you see FOSS as<br />

a viable option?<br />

7. In your opinion in what areas would FOSS be more advantages to you?<br />

References<br />

Chan, A. S. (2007) Retiring the Network Spokesman, Science Studies 20 :78-99.<br />

Dedrick, J. and West, J. (2003) A Grounded Theory of Innovation and Standards Adoption,” Proceedings of the<br />

Workshop on Standard Making: A Critical Research Frontier for Information Systems (MISQ Special Issue<br />

Workshop):12-14<br />

Delone, W.H. and McLean, E.R (2003) Journal of Management Information Systems 19 (4):9-30.<br />

Gable, Guy, Sedera, Darshana, Chan and Taizan (2003) Enterprise systems success: a measurement model, Proceeding<br />

of the 24th International <strong>Conference</strong> on Information Systems, Seattle, USA:576-591.<br />

Hanna N. K. (2006) From Envisioning to Designing e-Development: The Experience of Sri Lanka, World Bank Publications:179-181.<br />

ICTA (2008) Re-engineering Government, accessed on 20-01-2008,<br />

<br />

Kenneth Wong (2004) Free/Open Source Software: Government Policy, Asia-Pacific Development Information Programme,<br />

ELSEVIER, accessed on 02-05-2008, <br />

Kwan, S. K and West, J. (2005) A conceptual model for enterprise adoption of open source software,” The Standards<br />

Edge: Open Season. Sheridan Books, Ann Arbor, Michigan<br />

Lemos, R. (2003) Merrill Lynch: Linux saves money, 7 June, Cnet News.com, accessed on 27-06-2008,<br />

<br />

Maguire, J. (2003) Windows vs. Linux: TCO Feud Rages On, Newsfactor Network, accessed on 27-06-2008,<br />

<br />

Orzech, D. (2002) Linux TCO: Less Than Half The Cost of Windows, CIO Update, accessed on 27-06-2008,<br />

<br />

Reddy, B. and Evans, D. (2002) Government Preferences for Promoting Open-Source Software: A Solution in<br />

Search of a Problem, Social Sciences Research Network, accessed on 01-05-2008,<br />

<br />

Schindler Esther (2008) Open Source is Entering the Enterprise Mainstream – Survey Shows, CIO.com, accessed<br />

on 27-06-2008,<br />

http://www.cio.com/article/375916/Open_Source_is_Entering_the_Enterprise_Mainstream_Survey_Shows><br />

Stoltz, M. (1999). The Case for Government Promotion of Open Source Software, NetAction (27), accessed on 05-<br />

05-2008, <br />

Wikipedia (2005) Free and open source software, accessed on 10-01-2008,< http://en.wikipedia.org/wiki/FOSS><br />

Wong Kenneth (2004) Free/Open Source Software:Government Policy, United Nations Development Programme –<br />

Asia Pacific Development Information Programme (UNDP-APDIP), ELSEVIER.<br />

Yin R.K. (2003) Case study research: design and methods, 3rd edition, Sage Publications Inc, USA.<br />

252


Evaluating the Success of Customer Relationship<br />

Management (CRM) Systems<br />

Farnoosh Khodakarami and Yolande Chan<br />

Queen’s University, Kingston, Canada<br />

Fkhodakarami@business.queensu.ca<br />

Ychan@business.queensu.ca<br />

Abstract: Based on the Delone and McLean IS success model (1992), this qualitative study explores customer<br />

relationship management (CRM) system success factors. A case research approach was applied to examine CRM<br />

applications in three organizations. Through these case studies, five factors of CRM success were explored, namely:<br />

system quality, customer information quality, system use, user satisfaction, and system impacts. With regard to<br />

system quality, a main determinant of success was the system’s primary characteristics. CRM systems can be<br />

categorized as operational, analytical and collaborative, based on their characteristics. The study showed how<br />

system’s characteristics directly affect the purpose of system use. Operational CRM systems are mostly used for<br />

providing customer services. Analytical systems are primarily used for decision support. Collaborative systems are<br />

used both for work integration and providing service for internal and external customers. The study showed that if<br />

systems properly address users’ expectations and employees have the required skills to work with systems,<br />

employees are more willing to utilize the capabilities of systems to support CRM processes and, in turn, are more<br />

satisfied with the outcomes. Customer information quality relates to system quality; integrated CRM systems that are<br />

based on standard platforms are more capable of generating high quality customer information in a timely manner.<br />

Individual and organizational impacts of CRM systems were explored. Individual productivity, improved decision<br />

making and planning, learning and awareness were shown to be important individual benefits that CRM systems<br />

provide for employees. At the organizational level, the study showed that CRM systems help organizations to acquire<br />

and share more knowledge about their customers, and improve business processes, products and services.<br />

Keywords: customer relationship management, CRM, system evaluation, IS success, system effectiveness<br />

1. Introduction<br />

In response to increasing competitiveness in the global market, organizations are looking for means to<br />

manage their relations with customers more effectively over time. As a result, customer relationship<br />

management (CRM) has become a central part of most businesses and many organizations are<br />

implementing CRM systems. CRM systems are a category of organizational information systems that<br />

enable organizations to provide services for customers, gather and analyze customer information and<br />

manage their relationships with customers more effectively. However, despite substantial investments,<br />

many CRM systems fail to meet expectations. Besides, organizations often do not know appropriate<br />

success criteria to use to evaluate their CRM systems’ effectiveness. This study addresses these issues<br />

by drawing on a qualitative case research approach.<br />

The rest of this paper is organized as follows. First, the theoretical backround and research framework<br />

are presented. Then, the research method and findings are described. Finally, conclusions, limitations<br />

and implications are discussed.<br />

2. Theoretical background<br />

The main objective of CRM is to retain current customers through increasing their loyalty and to select<br />

new customers that provide higher profitability (Hansotia 2002). With the growth of information<br />

technology applications in organizations, CRM systems are popular. Successful CRM systems help<br />

organizations to manage customer information more effectively and increase customer satisfaction,<br />

customer loyalty and customer profitability through meaningful communication with customers (Winer<br />

2001). CRM systems fall mainly into three classes (Bose and Sugumaran 2003; Henning et al. 2003):<br />

Operational CRM systems that aim at automation of CRM processes to improve efficiency and<br />

productivity. Examples of such systems are sales and marketing automation systems (e.g., point of<br />

sales (POS) systems), call support centres and customer databases.<br />

Analytical CRM systems that aim to provide better understanding of individual customers’ behaviours<br />

and needs. Analytical systems incorporate various tools for consumer behaviour predictive modeling<br />

and purchase pattern recognition, such as data mining, data warehouse and OLAP (online analytical<br />

processing).<br />

253


Farnoosh Khodakarami and Yolande Chan<br />

Collaborative CRM systems that manage and integrate customer communication channels and<br />

networks. Email, websites, internal portals, and tele-, video- and web-conferencing applications are<br />

examples of collaborative systems that organizations use.<br />

Organizations invest in CRM systems and in turn expect the systems to positively impact their<br />

organizations and various customer related processes (Barua et al. 1995; King and Burgess 2008).<br />

Accurate measuring of the impacts of CRM systems requires “understanding systems’<br />

multidimensionality, and the development of a correspondent, standardized, validated, and robust<br />

measurement instrument” (Sedera and Gable 2004, p. 449).<br />

However, within the CRM domain, there is a lack of research that evaluates the success of CRM systems<br />

in actual use. To measure the success of information systems, Delone and McLean (1992) developed a<br />

model that includes both individual and organizational impacts of information systems. Considering the<br />

CRM systems as a category of organizational information systems, their success can be evaluated using<br />

the Delone and McLean model.<br />

This model (see Figure 1) has been widely cited and has been a valuable means to measure IS success<br />

in organizations (Gable, Sedera and Chan 2008).<br />

Figure 1: DeLone and McLean IS success model (1992)<br />

In the IS success model, two main factors, system quality and information quality, influence system<br />

usage and user satisfaction, which in turn lead to individual and organizational impacts. System quality<br />

refers to the desirable characteristics of an information system, such as system flexibility and ease of<br />

learning, while information quality refers to the desirable characteristics of the system outputs, such as<br />

management reports and Web pages. ”Use” can have different meanings. It can refer to the amount of<br />

use, frequency of use, and extent of use, as well as the nature and purpose of use (Petter et al. 2008).<br />

Examples of measures in each category (i.e., system quality, information quality, individual and<br />

organizational impacts) were gathered in a multidimensional IS success measures instrument developed<br />

by Sedera and Gable (2004) that also helped to inform our study. See Table 1.<br />

Table 1: Validated measures for IS success, Sedera and Gable (2004)<br />

For the purpose of this exploratory study, a framework was developed based on the Delone and McLean<br />

(1992) model. This framework is discussed in the next section.<br />

254


3. Research framework<br />

Farnoosh Khodakarami and Yolande Chan<br />

The impacts of information systems are often interdependent and influenced by various factors. The<br />

research framework (see Figure 2) indicates the factors that are investigated in this research. Not all of<br />

the DeLone and Mclean factors are considered in this paper because of space constraints. Some of the<br />

most important variables that affect the success of CRM systems are as follows:<br />

Figure 2: Research framework<br />

3.1 System quality<br />

System quality refers to the desirable characteristics of an information system (Petter et al. 2008).<br />

System quality is a multi-dimentional measure. Many researchers focus on the engineering<br />

characteristics and features of the system as main indicators of system quality (Ravichandran and Rai<br />

2000). Therefore system features such as integration, complexity and flexibility have been considered as<br />

system quality measures in a broad range of IS sucess studies (Petter et al. 2008). Besides, in regard to<br />

CRM systems, the primary characteristics of the system (i.e., operational, analytical and collaborative<br />

characteristics) can be considered as system quality factors. It is worthwhile to investigate how these<br />

characteristics affect systems’ success. In addition, users’ judgement is also critical for the evaluation of<br />

system quality. Measures like user requirements, ease of use and ease of learning reflect the users’<br />

evaluation of system quality (Ravichandran and Rai 2000, Petter et al. 2008).<br />

3.2 Customer information quality<br />

In the DeLone and McLean (1992) model, information quality refers to the desirable characteristics of the<br />

system output, namely the quality of information that system produces in the forms of reports, analyses<br />

and web pages (Petter et al. 2008). The main output of the CRM system is customer information, and<br />

effective management of customer information is a central characteristic of CRM systems; it helps<br />

organizations to better understand their customer needs and expectations and to respond to those needs<br />

in a timely manner (Hansotia 2002). Large amounts of data available to organizations through different<br />

communication channels are becoming a challenge for organizations. CRM success is therefore strongly<br />

associated with systems’ capability to manage the data and provide real time, meaningful, accurate and<br />

high quality customer information (Reid and Catterall 2005). Thus, availability of customer information is<br />

in fact a good indicator of the quality of the information.<br />

3.3 System use<br />

System use is an important dimension of IS success and is associated with the impact or benefits<br />

realized from IS (Petter et al. 2008). It indicates the degree and manner in which employees utilize the<br />

capabilities of an information system, and it can be described as amount of use, frequency of use and<br />

appropriateness of use (DeLone and McLean 2003; Petter et al. 2008). Most studies have measured the<br />

extent of use ( i.e., amount and frequency of use). However, the extent of use may not be the best<br />

measure of IS use since more use is not always better. To overcome this deficiency, Doll and Torkzadeh<br />

(1998) suggested that system use can be measured based on the purpose of use. They identified<br />

decision support, customer service and work integration as three main usage behaviours. Decision<br />

support refers to the extent to which the information system is used to analyze data and make decisions.<br />

Customer service is the extent to which the information system is used to provide service for both internal<br />

customers (employees) and external customers. Work integration incorporates horizontal and vertical<br />

255


Farnoosh Khodakarami and Yolande Chan<br />

integration. Horizontal integration refers to the extent to which the information system facilitates<br />

communication in work groups and helps employees to collaborate on various tasks. Vertical integration<br />

refers to the extent to which the system is used to “plan one’s work, monitor performance and<br />

communicate vertically with superiors and subordinates” (Doll and Torkzadeh 1998, p. 174).<br />

3.4 User satisfaction<br />

User satisfaction refers to the user’s level of satisfaction with system reports, outcomes and support<br />

services (Delone and McLean 2003). Scholars use various measures for user satisfaction. End User<br />

Computing Satisfaction (EUCS) (Doll and Torkzadeh’s 1994) and the User Information Satisfaction (UIS)<br />

instrument (Ives et al. 1983) have been the most widely used instruments. However, some scholars use a<br />

single measure for user satisfaction (e.g., Rai et al. 2002) and others measure positive user attitude as<br />

an indicator of IS success (e.g., Coombs et al. 2001). While some measures of user satisfaction focus on<br />

cognitive aspects of user satisfaction and the "extent to which users believe the information system<br />

available to them meets their information requirements" (Ives et al. 1983, p. 785), other measures focus<br />

on affective responses (i.e., favorable and unfavorable attitudes toward systems) as indicators of user<br />

satisfaction (Melone 1990). In this research both cognitive and affective responses are explored.<br />

3.5 System impacts<br />

System impacts refers to “the extent to which IS are contributing to the success of individuals, groups,<br />

organizations, and industries” (Petter et al. 2008, p.239). CRM systems provide benefits at different<br />

organizational levels. Individuals in different departments use CRM systems to accomplish their tasks<br />

more effectively and efficiently. Individual impacts indicate the extent to which CRM has influenced the<br />

capabilities and effectiveness of individual users. Organizations also take advantage of CRM systems.<br />

Organizational impacts indicate the extent to which CRM has promoted improvement in organizational<br />

results and capabilities (Gable et al. 2008). The Petter et al. (2008) review of literature on IS success<br />

revealed that few studies have considered organizational impacts of information systems. However,<br />

despite measurement difficulties, impacts of information systems cannot be monitored only at the<br />

individual level. Specifically, in the context of CRM, we know these systems provide benefits beyond the<br />

individual level. Within organizations, several departments such as marketing, sales, customer relations<br />

and R&D can substantially take advantage of CRM systems. Business process change, overall<br />

productivity, increased outcomes/outputs are some examples of organizational benefits of IS.<br />

4. Research method<br />

This study explores CRM system usage and evaluates systems’ effectiveness in organizational contexts.<br />

Given the exploratory nature of this study, a qualitative case study approach was used. The case studies<br />

allowed the researchers to learn about CRM systems in their business contexts and explore various<br />

factors that affected the usefulness of the systems. Case studies are specifically appropriate for<br />

information systems research; due to the rapid pace of information systems evolution, previous studies of<br />

the phenomena are limited. Case studies provide researchers with valuable insights to study phenomena<br />

and generate theories from practice (Benbasat et al. 1987). Three Canadian organizations were selected<br />

for this study. These organizations represent different industries and vary in terms of CRM capabilities,<br />

system development levels and organizational characteristics (see Table 2). The organizations use CRM<br />

systems and other information systems to support customer relations and processes. Call centres, POS<br />

and customer databases are the main operational systems used. For customer analysis, the electronics<br />

organization uses data warehouse and data mining systems, but the other two organizations largely rely<br />

on Excel analyses. In regard to collaborative systems, e-mail, company websites and social media (e.g.,<br />

Facebook, online communities and forums) are the main means of communication with customers. Each<br />

organization has several internal portals to support customer-related processes. The electronics<br />

organization has an e-support website to provide online customer services.<br />

During a four-month period in 2010, twelve telephone and in-person interviews were conducted with<br />

CRM users and with managers and employees from customer service, marketing, and IT departments<br />

within the three organizations. Sample interview questions are presented in Appendix. Interviews were<br />

recorded, transcribed and coded using the Nvivo program. The results are discussed in the next section.<br />

This research methodology is rigorous, satisfying criteria that Dube and Pare (2003) have recommended<br />

for a good IS positivist case study. Criteria for research design include: clear research questions, a priori<br />

specification of the constructs, a clean theoretical slate, and a multiple-case design. As recommended by<br />

Dube and Pare (2003), to assist with data analysis and the formulation of arguments, a case study<br />

256


Farnoosh Khodakarami and Yolande Chan<br />

database was developed as interviews were coded, field notes recorded, cross-case patterns explored,<br />

and quotes tracked.<br />

Table 2: Comparison of case characteristics<br />

Organization<br />

Organization Size<br />

Product/Service System Development Level<br />

Structure<br />

Electronics Large<br />

Health Medium<br />

Several companies,<br />

Local stores across<br />

Canada<br />

One company, Local<br />

stores across<br />

Canada<br />

Education Small One local school<br />

Various categories<br />

of electronic<br />

products<br />

Weight loss<br />

services, health<br />

and nutrition<br />

products<br />

Services<br />

(Educational<br />

programs)<br />

Well developed systems with various<br />

applications, systems are upgraded<br />

regularly<br />

Few systems that were recently<br />

launched, system development is in<br />

progress<br />

Few systems that have been used for<br />

years with minor upgrades, systems<br />

applications are not well utilized<br />

5. Results and discussion<br />

Using the case data, the relationships among system success factors were explored. Interviewees were<br />

asked to explain the benefits that systems provide for them and their organizations. Figure 3 presents the<br />

relationships that were identified through the interviews. These relationships are discussed below.<br />

Figure 3: Summary of findings<br />

257


System quality: Customer information quality<br />

Farnoosh Khodakarami and Yolande Chan<br />

System quality affects customer information quality in several ways. An important factor that determines<br />

the quality of systems is system integration. The interviews revealed that system integration greatly<br />

affects the availability of customer information. Customer information exchange is an important issue for<br />

many organizations; integrated systems facilitate the flow of information through the organization and<br />

allow the employees to get access to the right information in a timely manner. In the health organization,<br />

the disintegration of systems has increased the work overload in s, since the employees have to spend<br />

time to manually gather and combine all the information they need to do analyses and make decisions.<br />

There’s still not that direct connect between the reports. And I have to go back through so<br />

much manual work to find out how many leads each centre had and then try to correlate it to<br />

how many actually showed up and then manually again to find out how many sold.<br />

Regional Director, Health Organization<br />

In addition to system integration, system flexibility affects the availability of customer information. Flexible<br />

CRM systems allow users to define their specific information requirements and get access to that<br />

information. It is assumed that more system flexibility is beneficial for the availability of information;<br />

however an interesting observation was that the high level of flexibility will in fact restrict the availability of<br />

information in some occasions. For instance, the CRM system which is used in the education<br />

organization has a high level of flexibility that allows the employees to easily create new databases.<br />

These databases consist of custom-made fields based on individuals’ preferences. As a result,<br />

paradoxically, employees cannot easily retrieve the information they need from other employees’<br />

databases and integrate it to their own databases and this restricts the flow of information.<br />

The amount of flexibility is incredible, almost too flexible. It allows us to keep creating these<br />

separate databases as often as we can without relating any information between them and<br />

that's a big challenge.<br />

IT Associate Director, Education Organization<br />

System quality: System use<br />

System quality affects the use of systems. Both purpose and frequency of use are affected by system<br />

quality. The data suggested that purpose of use is associated with the primary characteristics of CRM<br />

systems. Within the studied organizations, operational CRM systems such as call centres and point of<br />

sales (POS) systems are mostly used for providing customer services for external clients. Analytical<br />

systems are primarily used for decision support. Employees use analytical systems (e.g., data<br />

warehouses, Excel) to analyze customer information and discover patterns of consumer behaviour.<br />

These analyses aid the decision making processes to a great extent. Collaborative systems have the<br />

highest applicability since they are used for multiple purposes. Some of these systems are used for<br />

vertical integration. For instance, in the health organization and the electronics organization, district<br />

managers have regular web conferencing and teleconferencing with the store managers to monitor<br />

stores’ performance and share customer experiences and best practices. A group of collaborative<br />

systems such as portals supports horizontal work integration. Through the internal portals, employees<br />

collaborate on projects and collectively accomplish tasks. Another group of collaborative systems<br />

facilitates internal and external customer service provision. All three organizations have e-learning<br />

websites that provide customer support training for employees. The electronics organization has a<br />

website that provides customer and product information that sales associates need. This website is<br />

updated regularly to include new products. This organization also has an e-support website that supports<br />

repair and maintenance services for external customers.<br />

There are other system quality variables that affect the frequency of use and employees’ willingness to<br />

use the systems. User requirements are one of these variables. If a system fails to address users’<br />

requirements properly, understandably employees are less likely to use it. This is particularly a challenge<br />

for the education organization. This organization uses a CRM system which was primarily developed to<br />

support sales and marketing processes. Therefore, the CRM system’s nomenclature and functionalities<br />

are not fully compatible with the particular requirements of an educational organization. As a result, the<br />

capabilities of the systems have been largely underutilized.<br />

258


Farnoosh Khodakarami and Yolande Chan<br />

A lot of examples in the training guide are sort of product related, which is not really the area<br />

that we fall under… if we go out into the more sales and marketing aspect of it [the CRM<br />

system] we might realize that it’s limited in what we’re trying to do. We’re just not using those<br />

parts.<br />

Director of Executive Development Centre, Education Organization<br />

Ease of learning is another system quality factor that influences system use. The interviews at the<br />

education organization revealed that the CRM system has sophisticated analytical capabilities and a<br />

database management tool that “scares people”. It is not easy to learn and utilize these capabilities since<br />

“it takes a different IT skill set and is not something that the average employee can do”. As a result,<br />

employees do not use these tools and merely rely on Excel spreadsheets to do basic analyses and<br />

reporting.<br />

Customer information quality: System use<br />

No direct relationship was found between customer information quality and system use. In the literature<br />

also there is little research demonstrating this relationship. The Petter et al. (2008) review of 180 papers<br />

on IS success showed that, at the time, only six papers had discussed this relationship. In general, Petter<br />

et al. (2008) did not find sufficient data to support this relationship.<br />

System quality: User satisfaction<br />

As expected, it became clear that if a system has the capability to meet the users’ requirements,<br />

employees are more satisfied with the system. For instance, the health organization has a system that<br />

allows employees to monitor clients’ weight loss progress, easily get access to their profiles and<br />

purchase behaviour, perform analyses, and provide consultancy services. This system was developed<br />

internally to address the specific requirements of the users; therefore, employees are very satisfied.<br />

My first real experience with [the new system] was just like there was a wild moment for me<br />

where I went 'Wow!' it gives me everything I need, I can actually be more effective as a<br />

regional manager by spending an effective administration day, reviewing what's going on in<br />

centres before I go there.<br />

Regional Director, Health Organization<br />

System integrity also affects user satisfaction. In many cases, lack of system integration leads to<br />

duplication in data entry and manual work in order to aggregate data from different sources; this leads to<br />

dissatisfaction. Similarly, if the CRM system is not compatible with other information systems, it can<br />

cause problems and unnecessarily increase the workload, which also leads to employee complaints.<br />

Customer information quality: User satisfaction<br />

Availability of customer information affects user satisfaction. CRM systems provide employees with the<br />

information they need for their analyses, planning and decision making processes. When CRM systems<br />

provide the right information in a timely manner, employees feel satisfied with the system. However, if the<br />

information is not properly organized, it will take a long time for employees to pull out the right information<br />

they need. This inefficient process can be tedious and time consuming, and lead to user dissatisfaction.<br />

We put a lot of information into different areas and therefore pulling out the information is<br />

difficult …. It can sometimes take a few hours to get information that if it was done properly<br />

could be probably at your fingertips.<br />

Director of Operations, Education Organization<br />

System use: Impacts<br />

CRM systems provide various benefits for organizations. Interviewees were asked to discuss some of the<br />

benefits that systems provide for their individual work and for the organization. In general, most of the<br />

interviewees admitted that CRM systems have improved their business capabilities and helped them<br />

know their customers better and make more effective decisions. The main systems’ impacts which were<br />

identified through the interviews are as follows:<br />

259


Farnoosh Khodakarami and Yolande Chan<br />

Individual impacts:<br />

Productivity<br />

Interviewees believed that CRM systems help them work more productively. They spend less time<br />

acquiring and aggregating information, they can perform more comprehensive analyses, and they provide<br />

better customer service. In addition, the feedback that organizations receive through various collaborative<br />

systems helps managers to evaluate employees’ performance and provide recommendations for training<br />

that can help employees to further improve their performance.<br />

Improved decision making and planning<br />

Systems also improve individual decision making and planning. Using the systems, employees have<br />

access to more knowledge resources and can get the information they need for decision making in a<br />

reasonable time. Exchanging ideas and feedback through the collaborative systems also helps managers<br />

with their decision-making and planning tasks.<br />

Learning<br />

Information systems provide learning opportunities for employees and customers. The organizations<br />

studied have e-learning systems and internal portals which provide documents and knowledge resources<br />

related to customers. The electronics organization also has an internal portal that provides product<br />

information for sales associates and call centre agents who can learn about the specific features of each<br />

product. This enables the sales associates and call centre agents to provide better customer service.<br />

Awareness<br />

There are several systems that increase employees’ awareness of events, market fluctuations and<br />

product changes. The health organization has an internal portal where all events and discussions are<br />

posted. Within the education organization, the instant messaging tool is broadly used for this purpose. In<br />

the electronics organization, agents benefit from increased awareness as a result of a “product alert”<br />

CRM feature. If a specific problem is identified regarding a new product, the call centre manager makes<br />

the call centre agents aware of this problem by assigning an alert to the product. It helps the agents to<br />

provide better service for customers when they have inquiries about the particular product.<br />

Organizational impacts<br />

Knowledge sharing and communication<br />

Collaborative CRM systems facilitate knowledge sharing and communication through the organization.<br />

For instance, in the electronics organization, employees can share their new ideas through an electronic<br />

suggestion box. Web-conferencing is also used in all three organizations to facilitate real-time<br />

communication in team meetings. Online communities and social media (e.g., Facebook and Twitter) are<br />

some of the systems that these organizations use to communicate with potential and actual customers.<br />

Process improvement<br />

Knowledge sharing and feedback from employees and customers help the organizations to improve<br />

various processes (e.g., marketing, advertising, and financial forecasting). The electronics organization<br />

has launched an advertising and product recommendation process which is supported by analytical CRM<br />

systems. By analyzing customer purchase history for each individual customer, the organization is able to<br />

suggest products that match each customer’s needs and complements his or her past purchases.<br />

Product /service improvement<br />

CRM systems, and specifically analytical CRM systems, help organizations to better understand<br />

customers’ need and predict their behaviour. Customer feedback which is collected through collaborative<br />

systems also reveals customers’ expectations. Having more detailed knowledge about customers helps<br />

the company to improve products and services. In addition, CRM systems enable organizations to<br />

provide a broader range of services to their customers. For instance, the health organization has<br />

launched an online system which allows customers to monitor their weight loss progress at any time,<br />

consult with health and nutrition experts, and receive supplemental resources and articles that give them<br />

the information they need to maintain their weight loss. The e-support system of the electronics<br />

organization is another effective system that helps to provide superior self-service support for customers.<br />

The system not only provides manuals and information on products but also guides customers to<br />

personally recognize product problems and fix them without the need to refer problems to repair shops.<br />

260


6. Conclusion<br />

Farnoosh Khodakarami and Yolande Chan<br />

Organizations are interested in evaluating the effectiveness of CRM systems; however the effectiveness<br />

of systems is influenced by various technological and organizational factors, and the measurement of<br />

systems’ success is challenging. Using the Delone and McLean (1992) multi-dimensional IS success<br />

model, this study explored factors that affect the success of CRM systems. The qualitative case studies<br />

provided the researchers with insights and real examples of how CRM systems are used in their<br />

organizational contexts and what variables contribute to their success. System integration was found to<br />

be an important technological variable that affects the use of systems. Systems integrity facilitates<br />

information sharing and allows employees to access accurate information in a timely manner. It was<br />

confirmed that if systems properly address the users’ requirements, they are more likely to be used.<br />

However, the level of systems’ complexity should match the employees’ IT skills; otherwise systems’<br />

capabilities will not be utilized.<br />

CRM systems help employess be more productive. They also provide learning opportunities for<br />

employees and customers. But CRM systems’ impacts go beyond the individual level. At the<br />

organizational level, systems facilitate communication and knowledge sharing throughout the<br />

organization and with customers. Furthermore, CRM systems enable organizations to provide a broader<br />

range of services for customers and improve products and business processes.<br />

In summary, CRM systems have major impacts at different levels and their effectiveness depends on<br />

several technological and organizational factors. As a result, in order to evaluate the effectiveness and<br />

success of CRM systems, organizations should draw on multi-dimensional measures and pay careful<br />

attention to the interdependency of these measures.<br />

7. Limitations<br />

Some study limitations should be acknowledged. Our findings are based on case studies and interviews<br />

in three industries. As a result, the generalizability of our findings is limited. However, our findings can<br />

serve as a starting point for other researchers exploring measures for the evaluation of CRM systems.<br />

We invite other researchers to conduct quantitative, large-scale follow-on explorations.<br />

8. Research implications<br />

This exploratory research contributes to our understanding of CRM system success measures. Previous<br />

studies on IS success have rarely focused on CRM systems. The success dimensions discussed in this<br />

paper can serve as the basis for a CRM system effectiveness construct. Future research can enhance<br />

and validate measures for the evaluation of CRM systems. The current study‘s findings can also help<br />

researchers and organizations better understand the strengths and weaknesses of current CRM systems,<br />

and collaborate on the design of future systems.<br />

9. Appendix 1: Sample interview questions<br />

On a daily basis, how often do you use CRM systems?<br />

For what purposes do you use CRM systems?<br />

What benefits do CRM systems provide for you and your organization?<br />

To what extent are CRM systems successfully and effectively supporting business processes within<br />

your department or in the whole organization?<br />

In your opinion, what are the strengths and weaknesses of your organization’s current CRM<br />

systems?<br />

What are your suggestions for improving CRM systems to match your requirements?<br />

In your opinion, what are the barriers for more extensive and effective use?<br />

How does your organization help you better utilize CRM applications?<br />

References<br />

Benbasat, I., Goldstein, D. K. and Mead, M. (1987) "The Case Research Strategy in Studies of Information Systems".<br />

MIS Quarterly, Vol 11, No.3, pp.369-386.<br />

Bose, R. and Sugumaran, V. (2003) "Application of Knowledge Management Technology in Customer Relationship<br />

Management". Knowledge and Process Management, Vol 10, No.1, pp. 3-17.<br />

261


Farnoosh Khodakarami and Yolande Chan<br />

Delone, W. and Mclean, E. (1992) "Information Systems Success: the Quest for the Dependent Variable".<br />

Information Systems Research, Vol 3, No.1, pp. 60–95.<br />

Doll, W. and Torkzadeh, G. (1994) "Developing a Multidimensional Measure of System-use in an Organizational<br />

Context". Informationand Management, Vol 33, No.4, pp.171–185.<br />

Dube, L. and Pare, G. (2003) "Rigor in Information Systems Positivist Case Research: Current Practices, Trends,<br />

and Recommendations". MIS Quarterly,Vol 27, No.4, pp. 597-635.<br />

Gable, G. G., Sedera, D., and Chan, T. (2008) "Re-conceptualizing Information System Success: The IS-Impact<br />

Measurement Model". Journal of the Association for Information Systems, Vol 9, No.7, pp. 377-408.<br />

Hansotia, B. (2002) "Gearing up for CRM: Antecedents to Successful Implementation". Journal of Database<br />

Marketing, Vol 10, No. 2, pp.121-132.<br />

Ives, B., Olson, M. and Baroudi, J. (1983) "The Measurement of User Satisfation". Communications of the ACM, Vol<br />

26, No.10, pp. 785–793.<br />

BIBLIOGRAPHY Melone, N. P. (1990) " A Theoretical Assessment of the User-Satisfaction Construct in Information<br />

Systems Research". Management Science, Vol. 36, No. 1, pp. 76-91 .<br />

Petter, S., DeLone, W., and McLean, E. (2008) "Measuring Information Systems: Models, Dimensions, measures,<br />

and interrelationships". <strong>European</strong> Journal of Information Systems, Vol 17, pp. 236–263.<br />

Rai, A., Lang, S. S. and Welker, R. B. (2002) "Assessing the Validity of IS Success Models: An Empirical Test and<br />

Theoretical Analysis". Information Systems Research, Vol 13, No.1, pp. 50-69.<br />

Ravichandran, T. and Rai, A. (2000) "Total Quality Management in Information Systems Development: Key<br />

Constructs and Relationships". Journal of Management Information Systems, Vol 16, No.3, pp.119-155.<br />

Sedera, D. and Gable, G. (2004) "A Factor and Structural Equation analysis of the Enterprise Systems Success<br />

Measurement Model", Twenty-Fifth International <strong>Conference</strong> on Information Systems, pp. 449-464, Association<br />

for Information Systems, Washington, USA<br />

Not all references cited in this article have been listed because of space constraints. Other references are available<br />

upon request from the authors.<br />

262


Toward a Novel Methodology for IT Strategic Planning<br />

Mouhsine Lakhdissi and Bouchaib Bounabat<br />

ENSIAS, Université Mohammed V, Morocco<br />

mlakhdissi@gmail.com<br />

bounabat@ensias.ma<br />

Abstract: Defining IT target state and the specific roadmap to attain it, are main focuses of IT strategic planning<br />

which is considered increasingly as the basis for IT decision-making and governance. Unfortunately, it hasn’t evolved<br />

with the same speed as other fields in the IT sphere. Most of the techniques, approaches and methods related to IT<br />

Strategic Planning date back to the 80s or 90s and are most often oriented business rather than IT. Enterprise<br />

Architecture is a really promising discipline aimed at capturing the as-is architecture of an enterprise, defining the<br />

target and the roadmap to get from existing to desired state. In that way, it is tightly related to IT strategic planning<br />

and it can provide a framework to fill the gap and contribute in structuring and formalizing IT Strategic Planning field.<br />

Enterprise Architecture benefits from a standardization effort as well as from tool support. Deliverables and artifacts<br />

are generally well defined and structured in the existing frameworks. This paper present a new methodology for IT<br />

Strategic Planning based on Enterprise Architecture.<br />

Keywords: IS strategic planning, enterprise architecture, methodology, metamodel, content framework<br />

1. Background<br />

1.1 Motivation<br />

With the complexity of today’s information systems and the necessity to make the existing IT assets more<br />

agile to provide for the constant business change, the task of governing and planning for IT assets<br />

become a key success factor for IT.<br />

Information Systems Strategic Planning is the discipline that deals with this task. Unfortunately, it hasn’t<br />

evolved with the same speed as other fields in the IT sphere. Most of the techniques, approaches and<br />

methods related to IT Strategic Planning date back to the 80s or 90s and are most often oriented<br />

business strategic planning rather than IT strategic planning (Hsu 1995). As a matter of fact, they don’t<br />

take into account the complexity of today’s information system and their diversity. Furthermore, this field<br />

lacks from a formal, rigorous and agreed upon methodology and suffers from the absence of tools to<br />

support, structure and industrialize the discipline<br />

Enterprise Architecture is a really promising discipline aimed at capturing the as-is architecture of an<br />

enterprise, defining the target and the roadmap to get from existing to desired state. In that way, it is<br />

tightly related to IT strategic planning and it can provide a framework to fill the gap and contribute in<br />

structuring and formalizing IT Strategic Planning field. Enterprise Architecture benefits from a<br />

standardization effort as well as from tool support. Deliverables and artifacts are generally well defined<br />

and structured in the existing frameworks.<br />

Existing Enterprise Architecture frameworks are of different types. While some frameworks like (Zachman<br />

1987) define a taxonomy for architecture artifacts, others like TOGAF (The Open Group 2009), tend to<br />

describe a process to produce architecture deliverables (Sessions 2007). The main concept underlying<br />

both the process and the taxonomy is the metamodel to describe architecture elements and to produce<br />

architecture deliverables.<br />

This metamodel is often either very poor to describe fully the architecture or not well structured to define<br />

the dependencies and the relationships between elements. We think that in order to define a more<br />

rigorous and structured methodology for IT Strategic Planning, it is necessary to define a rich and<br />

structured metamodel covering both architecture elements (processes, applications, data..) and<br />

transformation elements (programs, projects, budgets). This metamodel is the main focus of our work<br />

aimed at defining a new methodology for IT Strategic Planning.<br />

Our aim in this paper is to demonstrate the insufficiencies, deficiencies and inconsistencies in existing IT<br />

strategic planning methods and show how a new methodology based in part on the Entreprise<br />

Architecture practice could be proposed to address these problems. The metamodel we project to define<br />

could be used as a platform for describing the architecture, evaluating it and defining the needed<br />

transformations and planning them in term of programs/projects.<br />

263


Mouhsine Lakhdissi and Bouchaib Bounabat<br />

This paper proposes a new methodology for IT Strategic Planning based on Enterprise Architecture. One<br />

of the causes of IT project failure is poor strategic planning for them. This methodology can address the<br />

insufficiencies in existing approaches by using a rich metamodel and by combining architecture<br />

description and project definition.<br />

The second section presents IS Strategic Planning and Enterprise Architecture, compares the two<br />

disciplines and tries to bridge the gap between them. The content framework and the underlying<br />

metamodel is introduced in the the third section as a way to combine ISSP and EA. The fourth section<br />

describes the suggested metamodel and content framework. A comparison is made in the fifth section<br />

with existing metamodels before presenting future work and directions.<br />

1.2 Basic concepts<br />

The term methodology is very often misused in research as well as in practice. Many methods or<br />

frameworks are labelled “methodology” while they are not. So what is a methodology? It is defined by<br />

(Merriam 2010) as “a body of methods, rules, and postulates employed by a discipline” and by (TEAF<br />

2000) as "A documented approach for performing activities in a coherent, consistent, accountable, and<br />

repeatable manner". A method in the other hand is defined by (Merriam 2010) “a way, technique or<br />

process of or for doing something”<br />

From these definitions we can conclude that a methodology could be a structured composition of<br />

methods, rules and principles. It has a larger and more abstract scope than a method.<br />

Confusion is also often made between a framework and a methodology. (FEAF 1999) defines a<br />

framework as "A logical structure for classifying and organizing complex information". A framework is<br />

more of a structure that gives guidelines on some discipline without being prescriptive and without<br />

defining a structured process to undertake the underlying tasks.<br />

The table below summarizes some of the differences between the three concepts:<br />

Table 1: Comparison of method, framework and methodology<br />

Method Methodology Framework<br />

Scope Specific General General<br />

Abstraction Concrete Abstract Abstract<br />

Nature Prescriptive Prescriptive Descriptive<br />

Composition Simple Composite Composite<br />

Process Optional Mandatory Optional<br />

Product description Poor Average Detailed<br />

2. IT strategic planning and enterprise architecture<br />

2.1 IT strategic planning<br />

Strategy is defined by Chandler as “The determination of the basic long-term goals and objectives of an<br />

enterprise and the adoption of courses of action and the allocation of resources necessary for carrying<br />

out these goals”1 and by porter as “The art to build durable and defendable competitive advantage” 2 .<br />

One of the most complete definitions was given by (Arnoldo 1996), “A fundamental framework for an<br />

organization to assert its vital continuity, while, at the same time, forcefully facilitating its adaptation to a<br />

changing environment.”<br />

For most of the definition, strategic planning is focused on three main questions:<br />

Where we are?<br />

Where we want to go?<br />

How to get there?<br />

IS Strategic planning has been defined by (Lederer and Sethi 1992) as the process of identifying a<br />

portfolio of applications/projects that can help an organization achieve its business strategy. Its focus is<br />

on defining the IT roadmap in term of key initiatives, projects and transformations to be made on the<br />

existing information system with two main intentions:<br />

How to align information systems with business needs and overall strategy?<br />

264


Mouhsine Lakhdissi and Bouchaib Bounabat<br />

How to use information technology to change and impact the business?<br />

Due to the complexity of today’s information systems and the diversity of enterprise’s technology<br />

approaches, many methods have been defined to structure the ISSP process and techniques have been<br />

defined to address some aspects of the discipline. (Vitale 1986) classifies ISSP methods into two<br />

categories:<br />

Impact methods : trying to make It help create a positive impact and drive the change of the business<br />

Alignment methods : where the main focus is on aligning IT to respond to business needs and to help<br />

achieve strategic goals<br />

Among the methods used in IT Strategic Planning we have Critical Success Factors (CSF) (Rockart<br />

1979) which could be considered as an impact and alignment method, Business Systems Planning (BSP)<br />

(Wiseman 1988), Porter’s Competitive Forces Model (Porter 1980), Porter’s Value Chain (Porter 1985),<br />

and Scenarios (Schwartz 1991). Methods can be grouped together to constitute a methodology.<br />

Methodologies used for ISSP include those of the CCTA (1988, 1999) and Boar (2001).<br />

Many IT vendors and consultancy organizations use proprietary methods and/or methodologies, some of<br />

which are adaptations of open source approaches. Examples are Arthur Andersen’s Method/1 and<br />

Coopers and Lybrand’s Summit (Lederer and Sethi 1988, Min et al. 1999). It is also well known that<br />

organizations often develop their own in-house methodologies, often based on open or proprietary<br />

methods or approaches (Earl 1993, Lederer and Sethi 1988).<br />

2.2 Enterprise architecture<br />

ISO/IEC 42010: 2007 defines ‘‘architecture’’ as: ‘‘the fundamental organization of a system, embodied in<br />

its components, their relationships to each other and the environment, and the principles governing its<br />

design and evolution.’’. The (Open Group 2009) defines it as:<br />

“A formal description of a system, or a detailed plan of the system at component level to guide its<br />

implementation”<br />

“The structure of components, their inter-relationships, and the principles and guidelines governing<br />

their design and evolution over time”<br />

An architecture is typically made up of:<br />

A picture of the current state<br />

A blueprint, vision or detailed description for the future<br />

A road-map on how to get there<br />

Enterprise Architecture appeared in the eighties thanks to John Zachman who introduced the framework<br />

that bears his name. This framework consists of taxonomy for producing architecture artifacts from<br />

different viewpoints and perspectives. As a matter of fact, Enterprise Architecture has been defined by<br />

(Zachman, 2005) as a "set of descriptive representations (i.e. ‘models’) that are relevant for describing an<br />

Enterprise such that it can be produced to management’s requirements (quality) and maintained over the<br />

period of its useful life".<br />

Several other frameworks appeared subsequently, most of them initiated by government bodies like<br />

TAFIM (Technical Architecture Framework for Information Management), DODAF, MODAF or FEAF<br />

especially due to the requirement of the Clinger-Cohen Act.<br />

IT consulting firms created their own EA frameworks, based on the feedback from projects they<br />

undertook. Gartner as well as Cap Gemini or Accenture have their own EA frameworks which could be<br />

more accurately considered as EA practices as stated by (Session 2007)<br />

The Open Group Architecture Framework (TOGAF) started with TAFIM and reproduced practices and<br />

techniques used in other framework to constitute an EA framework of reference in the IT industry.<br />

TOGAF is with Zachman the two most used EA frameworks according to (Schekkerman 2005).<br />

TOGAF consist of:<br />

A architecture development methodology describing the process<br />

A set of guidelines and techniques supporting the methodology<br />

265


Mouhsine Lakhdissi and Bouchaib Bounabat<br />

A content framework with a metamodel describing the products (deliverables)<br />

Reference models that provide best practices to compare with<br />

A structure and description of the architecture repository (enterprise continuum)<br />

A capability framework for architecture governance and implementation<br />

TOGAF could be used in combination with Zachman where TOGAF defines the process and Zachman<br />

the deliverables. (Archimate 2009) defines a notation for architecture elements but also defines its own<br />

metamodel for architecture description.<br />

Enterprise Architecture could be used for different needs and in various contexts. It can operate as:<br />

A method to describe the enterprise as a whole with different levels and views of enterprise elements<br />

and their relationships. In this way it relates to Enterprise Modeling as was stated by (Lillehagen et<br />

al., 2005)<br />

A way to align the IS environment with the business reality and the strategic goals or to assess this<br />

alignment as described by (Bounabat 2006) (Elhari 2010)<br />

A modelling structure to define the vision for IS evolution or to describe in detail the IS to-be state<br />

A process to plan the migration between the as-is situation and the to-be state.<br />

All these EA use cases could be related to a step or a phase in term of process or deliverables of ISSP.<br />

2.3 Comparison and correlation<br />

A theoretical comparison of IT Strategic Planning and Enterprise Architecture was conducted by (Wilton<br />

2007) and (Beveridge and Perks 2003). These comparisons concluded that both ISSP and EA share the<br />

same intent and scope. (Wilton 2008) gave a more empirical comparison based on a survey which led to<br />

establishing a significant correlation between the two activities in term of topics they cover.<br />

The main difference that was highlighted by (Wilton 2008) is that ISSP tends to be process-oriented with<br />

little specification of deliverables and content while EA is product-oriented in that it defines the way the<br />

as-is and to-be state are described and modelled.<br />

We think that this difference tends to disappear due to the progress made in the field of Enterprise<br />

Architecture. As a matter of fact, with frameworks like TOGAF the gap is being bridged with a detailed<br />

process to produce architecture deliverables.<br />

Furthermore, we think that other differences are to be considered. They are summed up in the table<br />

below<br />

Table 2: Comparison enterprise architecture and IS strategic planning<br />

IS Strategic Planning Enterprise Architecture<br />

Period of development 1970-1995 1987-2010<br />

Process Linear Iterative and incremental<br />

Nature Business oriented Business and IT oriented<br />

Main deliverables Programs, projects, budgets Architecture description (as-is and<br />

to-be) and Roadmap<br />

Strategy support High Medium to low<br />

Techniques and modeling Techniques Techniques, Modeling<br />

Repository No Yes<br />

Tool support Rare Mainstream<br />

One of the main differences that still exists and that is related to the Enterprise Architecture practice is<br />

the fact that there is no concrete link between the architecture description and the programs/projects<br />

defined in the roadmap. This lack of correlation makes it difficult to address the strategic planning main<br />

objective which is planning for IT transformations with existing Enterrise Architecture frameworks.<br />

2.4 Bridging concept: Transformation<br />

A project is defined by (PMBoK) as “A project is a temporary endeavor undertaken to create a unique<br />

product, service or result.”. This definition underlines the fact that a project is intended to create a<br />

266


Mouhsine Lakhdissi and Bouchaib Bounabat<br />

product/service or result without mentioning the elements that the project will impact whether they are<br />

new elements created or existing element transformed.<br />

A project –in the context of ISSP and EA- could be defined as a set of transformations (including<br />

creations) applied on architecture elements. These elements could be business elements, application<br />

elements, data elements or technical elements or a combination of them.<br />

Elements are the basic constituents of architecture like applications, processes, servers, databases...etc.<br />

These elements are combined to create architecture models and diagrams. The transformations of these<br />

elements are combined as well to create ISSP’s projects and programs.<br />

3. Importance of a content framework and a metamodel<br />

The ISO/IEC 42010: 2007 definition of architecture as “the fundamental organization of a system,<br />

embodied in its components, their relationships to each other and the environment, and the principles<br />

governing its design and evolution.’’ highlight unequivocally the importance of the organization of<br />

elements and their relationships. This structure is defined through a metamodel of architecture elements.<br />

Enterprise Architecture is supposed to produce architecture artifacts. These artifacts are based on an<br />

architecture content framework as defined by TOGAF or an architecture map.<br />

3.1 Content framework<br />

The content framework defines the layers, views, questions and aspects that architecture description<br />

deals with. The importance of this framework is that it organizes, classifies and links architecture<br />

elements and artifacts. It is also interesting because it ensure the coherence and exhaustively of the<br />

metamodel.<br />

The content framework is classically defined as a bi-dimensional grid with lines representing layers or<br />

views and columns representing concerns and classifications.<br />

The content framework defines elements of the metamodel in a high level way emphasizing the global<br />

structure rather than the detail model.<br />

3.2 Metamodel<br />

The metamodel is the backbone of architecture description and methodology. The metamodel<br />

guarantees the exhaustiveness of overall architecture work and the coherence and alignment of<br />

architecture layers.<br />

It is similar in form to a Conceptual Data Model or a Class Diagram in UML. It is important in term of<br />

objects definition, attributes definition and relationships.<br />

Objects definition ensures the exhaustiveness and coverage of aspects like standardization,<br />

requirements and integration.<br />

Attributes provide the way to perform diagnosis and analysis on existing and future assets. Attribute<br />

can also cover aspects like security and performance necessary to the evaluation process.<br />

Relationships are very important to perform Gap Analysis inside the same layer and for alignment<br />

needs between layers.<br />

4. Existing content frameworks and metamodels<br />

Many metamodels have been defined explicitely or implicitely by EA frameworks. They are of different<br />

natures and focus depending on their intent. Some of them are poor in term of business or IS content.<br />

Others don’t take into account some aspect tightly related to EA and ISSP like :<br />

Requirements<br />

Strategy<br />

Standards<br />

Program and projects<br />

267


4.1 TOGAF metamodel<br />

Mouhsine Lakhdissi and Bouchaib Bounabat<br />

TOGAF metamodel which has been introduced only in the last version of the framework, is a theoretical<br />

metamodel somehow poor especially in the application, data and infrastructure layers. The metamodel of<br />

TOGAF describe elements of the four layers with their relationships and dependencies. The business<br />

layer is otherwise well detailed and provide for a good foundation for our metamodel. In the business<br />

layer nevertheless, only the static information is represented (no information flow) and the link between<br />

organization and data is not defined.<br />

We think that one of the contributions of this metamodel is its extensibility since TOGAF defines a core<br />

metamodel and many extensions applicable depending on the organization and project nature.<br />

Furthermore, some aspects like requirements, standards and project are not taken into consideration in<br />

the metamodel even if they are core in the framework<br />

4.2 Archimate<br />

Archimate is a structure and notation for architecture work. If used as a notation, it can be considered as<br />

complementary to frameworks like TOGAF or Zachman. Otherwise Archimate defines its proper content<br />

structure and metamodel. The structure is composed of three layers: Business, Application and<br />

Technology, with three views: passive structure, behavior and active structure.<br />

Archimate defines in a more detailed manner than TOGAF, the architecture elements of the three layers.<br />

In the business layer, it’s quite complete while still lacking definition of dynamic information flow and<br />

relationship between organization and information.<br />

In the application layer, it introduces integration elements like “application interface” and “application<br />

collaboration” but lack the underlying data integration elements (data flows, data messages…)<br />

One of the advantages of Archimate is its well defined structure combined with the detailed notation. It<br />

has intentionally -as stated in the specification document- omitted the strategy, standards and principles<br />

and strategic planning aspect. Other aspects like security and performance are also absent. They will add<br />

some of this aspect in future versions or as optional extensions.<br />

4.3 Zachman<br />

Zachman doesn’t define explicitly a metamodel. The metamodel is an implicit component appreciable<br />

through the models and artifacts defined in the Zachman grid.<br />

Through the different viewpoints and question, we can conclude that the Zachman metamodel has a<br />

large scope with detailed business and IT description. The main problem of Zachman is that it doesn’t<br />

define the relationship and dependencies between artifacts and thus between architecture elements<br />

which is a major drawback.<br />

Besides, Zachman doesn’t take into consideration Strategic Planning elements and their link with<br />

architecture elements.<br />

4.4 EA tools metamodel<br />

EA modeling tools have their own metamodel which are more often extensible. The architecture<br />

description is more or less detailed depending on the vendor positionning and history. Tools like Mega<br />

and Aris have more focus on the business layer with poorer IS capabilities, while tools like Power<br />

Designer are more relevant for IS and infrastructure description than in business modeling.<br />

Even though ISSP elements like projects can be defined and described, there is no link between the<br />

projects and architecture elements they impact.<br />

268


Mouhsine Lakhdissi and Bouchaib Bounabat<br />

5. Presentation of a new content framework and metamodel<br />

5.1 New content framework: NEOXIA Architecture Map (NAM)<br />

We introduce here a new content framework that is based on feedback from consulting projects on EA<br />

and ISSP. We call it NEOXIA Architecture Map. In this content framework we differentiate between<br />

Static element: tending to describe an element in a static most often hierarchical way<br />

Dynamic element : focusing on the dynamic view of the same element<br />

Figure 1: New content framework: Neoxia Architecture Map (NAM)<br />

We also distinguish between three natures of element:<br />

Structure elements: like organization and network<br />

269


Function elements: like services or functions<br />

Content element: like data or storage<br />

Mouhsine Lakhdissi and Bouchaib Bounabat<br />

Two additional but crucial architecture concerns emerge as intermediate planes between these elements:<br />

Security and integration.<br />

Figure 2: Architecture concerns in Neoxia Architecture Map (NAM)<br />

Between every layer and the one below we have an intermediate layer that is not shown in figure 1 but<br />

that is important to detail in some cases.<br />

As a matter of fact there is:<br />

A functional architecture layer between Business and the IS Architecture,<br />

A solution and development architecture layer between the IS and Software Architecture,<br />

A deployment and production architecture layer between the software and hardware layer.<br />

5.2 New Metamodel: NEOXIA Content Metamodel (NCM)<br />

The content metamodel is the mechanism by which we suggest to mix architecture and strategic planning<br />

element base on transformations. The content metamodel follows the overall structure of the content<br />

framework and could be illustrated as in figure 3.<br />

The suggested metamodel is composed of six layers:<br />

Strategy<br />

Requirements<br />

Business<br />

Information Systems<br />

Technology<br />

Strategic planning<br />

All layers are interrelated with static and dynamic element of the three natures: function, structure and<br />

content. Every layer is connected with the layer below with a realization link. A process is automated in<br />

an application which uses a database and are both deployed in a server. This dependence is<br />

fundamental to align the IS with the Business Architecture and the Technology with the IS Architecture.<br />

This link allows us also to analyze the gap between layers in term of coverage to make it possible to fill<br />

this gap in the strategic plan.<br />

The metamodel could be also represented as package and class diagrams. We focus on the strategic<br />

planning layer.<br />

The central concept is “Transformation” which is a migration from an as-is state to a to-be state of an<br />

architecture element. An architecture element could be any architecture object of the metamodel (ex:<br />

process, application, Hardware server…etc).<br />

270


Mouhsine Lakhdissi and Bouchaib Bounabat<br />

Figure 3: New Metamodel : NEOXIA Content Metamodel (NCM)<br />

is composed of<br />

defines<br />

is a set of<br />

originates from<br />

Program Project Transformation<br />

execute<br />

I nt er n al<br />

Or g a ni z ati o n<br />

Uni t : 2<br />

Impact<br />

Requirement<br />

originates from<br />

Strategic Objective : 2<br />

Figure 4: Class diagram of the package “IT Strategic Planning”<br />

originates from<br />

Gap<br />

concerns<br />

concerns<br />

Architecture Element<br />

A transformation is operated either:<br />

A realization of a strategic objective : this allows us to align the to-be IS situation with the strategy<br />

and to justify the strategic plan investments<br />

271


Mouhsine Lakhdissi and Bouchaib Bounabat<br />

A consequence of an IT or business requirement (principle, standard, rule, constraint) defined by the<br />

organization<br />

A result of a gap analysis: in that case the gap is observed on one or more architecture elements and<br />

the transformation is a way to fill the gap.<br />

We have as a result three types of transformations :<br />

Alignement transformation<br />

Requirement transformation<br />

Gap transformation<br />

Many transformations are grouped into projects which are managed through programs.<br />

6. Comparison with existing content framework and meta models<br />

The suggested metamodel defines the detailed structure and relations of architecture elements starting<br />

from strategy and requirement and going through the different levels (Business, Information Systems and<br />

Infrastructure) with the necessary link with Strategic Planning element like gap analysis, program and<br />

project.<br />

In addition of giving a more detailed structure for IS and infrastructure levels which are often poorly<br />

defined, the main contribution of this metamodel is the link it establishes between architecture elements<br />

and strategic planning elements.<br />

We summarize a theoretical comparison of our metamodel with existing metamodels presented in section<br />

3.<br />

Table 3: Comparison of the new metamodel with existing ones<br />

New metamodel TOGAF Archimate Zachman EA Tools<br />

NCM<br />

metamodels<br />

Requirement Yes No No.<br />

Partially Yes<br />

Undergoing<br />

Strategy Yes Yes No Yes No<br />

IT Planning Yes No No No Yes. To some<br />

Link between<br />

projects and<br />

architecture<br />

Yes No No No<br />

extent<br />

No<br />

Standards Yes No No No Yes. (Not<br />

native)<br />

Strategy definition Yes Yes No Yes Yes<br />

Business<br />

Detailed Detailed Detailed Average Poor<br />

description<br />

IS description Detailed Poor Average Detailed Detailed<br />

(Depending on<br />

Infrastructure<br />

description<br />

tool)<br />

Detailed Poor Average Detailed Detailed<br />

(Depending on<br />

tool)<br />

Tool support No Partial Yes Partial Yes<br />

Methodology<br />

support<br />

No. Undergoing Yes Yes No No<br />

Independence Yes Yes Yes Yes No<br />

7. Conclusion and future work<br />

We think that based on this proposed metamodel, a new methodology could be defined to cope with the<br />

needs of ISSP and to complement and enrich existing EA metamodels. The metamodel described was<br />

already used successfully in consulting projects in the public and private sector and was able to capture<br />

more meticulously architecture element and to support the process of IS Strategic Planning.<br />

The model could be enriched to highlight crossover architecture aspects like security, performance and<br />

integration. These aspects are very important in evaluating existing IT assets and in defining their target<br />

state.<br />

272


Mouhsine Lakhdissi and Bouchaib Bounabat<br />

A planned continuation of this work is to develop a basic modeling tool (or adapt an existing one) based<br />

on this metamodel and content framework with support to Enterprise Architecture as well as IS Strategic<br />

Planning techniques and activities. The tool will allow to put into practice the metamodel and to<br />

demonstrate the added-value of the methodology.<br />

Another extension is to formalize diagnosis and evaluation techniques into the meta model to make sure<br />

the whole IS Strategic Planning process is automated.<br />

References<br />

Archimate (2009), “Archimate 1.0 Specifications “, http://www.opengroup.org/architecture/togaf8-doc/arch/<br />

Arnoldo C. Hax (1996), “Strategy Concept and Process: A Pragmatic Approach” Barnes and Noble, bn.com<br />

Boar, B. (2001). The Art of Strategic Planning for Information Technology, John Wiley and Sons, New York, NY.<br />

Bounabat, B. (2006), “Enterprise Architecture Based Metrics for Assessing IT Strategic Alignment”, <strong>European</strong><br />

<strong>Conference</strong> On Information Technology Evaluation (ECITE), 2006.<br />

CCTA (1999). IS Strategy: process and products, Format Publishing Limited, Norwich.<br />

Chandler, Alfred A., Jr. (1962), Strategy and Structure: Chapters in the History ofAmerican Industrial Enterprise, The<br />

MIT Press, Cambridge, MA<br />

DODAF, U.S. DoD (2003). DoD Architectural Framework Version 1.0,<br />

http://www.teao.saic.com/jfcom/ier/documents/DOD_architecture_framework_volume1.doc,<br />

Elhari K., Bounabat. B (2010), “Strategic Alignement Assessement Based on Enterprise Architecture”, ICIME 2010<br />

FEAF, 1999, “Federal Enterprise Architecture Framework Specification”, http://www.whitehouse.gov/omb/e-gov/fea/<br />

Lillehagen F., Karlsen D., Enterprise Architectures – Survey of Practices and Initiatives, Rapport Computas 2005.<br />

Lederer, A. L. and Sethi, V. (1988). The Implementation of Information Systems Planning Methodologies, MIS<br />

Quarterly, September 1988, 445-461.<br />

Merriam Webster Dictionnary (2010), Dictionnary definition of Methodology and Method http://www.merriamwebster.com/dictionary/method,http://www.merriamwebster.com/dictionary/methodology?show=0&t=1290517042<br />

Pant S., Hsu C. (1995), Strategic Information Systems Planning: A Review, 1995 Information Resources<br />

Management Association International <strong>Conference</strong>.<br />

Perks, C. and Beveridge, T. (2003). Guide to Enterprise IT Architecture, Springer-Verlag, New York.<br />

Porter, M. E. (1985). Competitive Advantage: Creating and Sustaining Superior Performance, Collier Macmillan, New<br />

York, N.Y.<br />

Schekkerman J. (2005), Trends in Enterprise Architecture 2005: How are Organizations Progressing? Institute For<br />

Enterprise Architecture Development (IFEAD) report<br />

Sessions R. (2007), “A Comparison of the Top Four Enterprise-Architecture Methodologies”,<br />

http://msdn.microsoft.com/en-us/library/bb466232.aspx<br />

TEAF, U.S. Department of the Treasury (2000). Treasury Enterprise Architecture Framework Version 1,<br />

http://www.software.org/pub/architecture/teaf.asp,<br />

The Open Group (2009). “TOGAF as an Enterprise Architecture Framework”,<br />

http://www.opengroup.org/architecture/togaf8-doc/arch/<br />

Vitale, M., Ives, B. and Beath, C. (1986), “Identifying Strategic Information Systems,” Proc. 7th Int’l Conf. Inf. Sys.,<br />

San Diego, December 1986, pp. 265-276.<br />

Wilton, D. (2001). The Relationship Between IT Strategic Planning and Enterprise Architectural Practice, Journal of<br />

Battlefield Technology, 1, 18-22.<br />

Wilton, D. (2007). The Relationship between IS Strategic Planning and Enterprise Architectural Practice: a Study in<br />

NZ Enterprises, Information Resources Management Association (IRMA), Vancouver.<br />

273


Applying Innovative Information System to Improve the<br />

Effects of Community Health Promotion<br />

Chih-Yu Lin 1 , Wen-Yu Ho 2 , Chia-Hsien Wen 3 and Hui-Mei Hsu 4<br />

1<br />

National Chung Cheng University, Chiayi, Taiwan<br />

2<br />

China Medical University Beigang Hospital, Yunlin, Taiwan<br />

3<br />

Providence University, Taichung, Taiwan<br />

4<br />

National Kaohsiung Normal University, Kaohsiung, Taiwan<br />

cylin@ccu.edu.tw<br />

shiaoyin0505@gmail.com<br />

chwen@pu.edu.tw<br />

hmhsu@nknu.edu.tw<br />

Abstract: “Ottawa Charta” indicates that the importance of strengthen community actions is to achieve better health<br />

for residents. For achieving this goal, the health promotion works through community organizations is the most<br />

effective way. There are two models to promote community health in Taiwan. One is to set up a single community<br />

health department within a healthcare organization. The other one is to set up a community care center in the<br />

community by civic or government organizations. However, there is a common drawback of lacking a suitable and<br />

effective information system to evaluate the two models and serve as a basis to adjust future service content and<br />

efficiency. This study establishes an Internet information platform with Interactive Response System (IRS) to<br />

effectively evaluate the two models of community health promotion and to improve the services to upgrade the<br />

management of community health promotion efficiency. The results showed that the average satisfaction score of the<br />

Internet information platform is greater than 4 which mean that the innovative information system in this study is in<br />

accordance with the needs of community health department users and acquires a high satisfaction. The average<br />

satisfaction score of IRS is greater than 4 with the factor of “perceived usefulness” scores 5 which means residents of<br />

the community care center show a higher acceptance of using IRS in learning health education materials than that of<br />

traditional one. IRS can also improve diseases perception of community residents.<br />

Keywords: information system to improve community health promotion, community health promotion effective,<br />

interactive response system (IRS)<br />

1. Introduction and background<br />

In 1995, the Taiwan government implemented the National Health Insurance (NHI) System to provide a<br />

sound healthcare service for the citizens. According to the report of Department of Health, cancer and<br />

chronic diseases were the most serious health threats for the citizens in 2006. 78% of the funds were<br />

spent on the treatment of patients with chronic diseases. Therefore, patients who suffer from chronic<br />

diseases or have needs of long term care are the financial heavy burdens of NHI.<br />

Most citizens do not visit the doctor’s unless they are affected by a disease; this is the nature of the human<br />

beings. “Ottawa Charta” indicates that the importance of strengthen community actions is to achieve better<br />

health for residents. For achieving this goal, the health promotion works through community organizations<br />

is the most effective way. Health promotion includes promoting the cognition of health of individuals and<br />

communities, altering their attitudes for changing their behaviors of life, and seeking the methods to<br />

improve their health (Squyres, 1985).<br />

Many countries have enforced laws to protect the health of their people. For instance, Australia has<br />

enforced the Health Promotion Act in 1995; United States of America passed The Model State Public<br />

Health Act in 2003. With the rapid progress of information technology (IT), many enterprises, schools,<br />

hospitals and other organizations have employed IT to achieve their strategies and goals.<br />

There are two models to promote community health in Taiwan. One is to set up a single community health<br />

department within a healthcare organization. The other one is to set up a community care center in the<br />

community by civic or government organizations. However, there is a common drawback of lacking a<br />

suitable and effective information system to evaluate the two models and serve as a basis to adjust future<br />

service content and efficiency.<br />

The objectives of this study are to develop the community health promotion system to collect and analyze<br />

the changes of health cognition and health status of residents by using interactive response system (IRS).<br />

The data also can help the healthcare providers to understand the health conditions of the communities.<br />

274


2. Community health promotion<br />

Chih-Yu Lin et al.<br />

According to the Ottawa Charter, health promotion is the process of enabling people to increase control<br />

over, to improve their health, and to reach a state of complete physical, mental and social well-being<br />

(Ottawa Charter, 1986 ). As proposed by the Ottawa Charter, many countries have established health<br />

promotion policy such as Japan. In order to encourage all citizens to be healthy, keep away from diseases<br />

and get health information from the community, Japan announced Healthy Japan 21 in 2000. This policy<br />

focused on healthy dietary habits, promotion of physical activities, and diagnostic tests (Yajima, Takano,<br />

Nakamura & Watanabe, 2001; Dorjsuren, 2008).<br />

There are two mechanisms for health management: one is initiative model, which the health organization<br />

and government takes the initiative for the citizen’s health promotion; second is passive model, which the<br />

citizens and community takes the initiative in promoting their own health. Regardless of which model, the<br />

same objective is the prevention of diseases occurrence. Therefore, the subjects of health promotion<br />

include individuals, groups and community residents. The strategies of health promotion include health<br />

education, behavior motives, organization reengineering, economic effectiveness, and policy changes to<br />

further influence the living environment and people’s behaviors and to improve health and quality of life<br />

ultimately.<br />

Hancock and Duhlru (1986) defined community health as: continuing to create and enhance the physical<br />

and social environment of the community, while strengthening the community resources so that people<br />

can support each other, the implementation of all its living, and the maximum potential. WHO (1994)<br />

defined a healthy community as: healthy communities by healthy people, healthy environment and healthy<br />

social integration as a whole should be able to continuously improve the physical and social environment,<br />

expand community resources, support each other, and maximize potential. In brief, community health<br />

promotion is "health education" and "health policy", from the government, other non-governmental<br />

volunteer organizations, community participation, and hoping every community health issues for<br />

themselves, by residents spontaneous reaction to address community health problems, thus contributing<br />

to the community's health.<br />

The WHO (1978) proposed the concept of "National Health" that includes six principles, "strengthening<br />

community participation" is one of them. In addition, Ottawa Charter (1986) indicated that "strengthen<br />

community action" is one of the five health promotion actions. Both statements display community health<br />

promotion can be achieved through strengthening community actions. Community health building is a way<br />

to achieve community health. The community is regarding as a living body, and the whole community<br />

health is improved continuously through a planned action to change the health behaviors and decisions of<br />

community residents, and environments.<br />

The use of information technology in healthcare service has grown exponentially over the last several<br />

years (Pagliari C, Sloan D, Gregor P et al., 2005). Several studies noted that information technology was<br />

employed to improve health and health education. Heilman et al. (2011) indicated that Wikipedia is a key<br />

tool for public health promotion because the Internet has become an important health information resource<br />

for patients and the general publics. In 2004, Wikipedia editors formed a group called WikiProject<br />

Medicine to coordinate and discuss the Wikipedia’s medical content. Today, it has become the dominant<br />

online reference work. (Heilman et al., 2011). The Wantland’s study addressed important claims the use of<br />

web-based interventions compared to non-web-based interventions. The results indicated that the<br />

specified behaviors of individuals were changed by using web-based interventions, such as increasing<br />

participation in healthcare or improving body shape perception (Wantland, Portillo, Holzemer, Slaughter, &<br />

McGhee, 2004 ).<br />

3. Methods<br />

3.1 Case background<br />

The scale of case hospital is 429 beds, and the employees include 51 doctors, 195 nurses, 49 other<br />

medical staff, and 125 administrative personnel. We established a large group in this hospital for<br />

community service, which has been further divided into 22 smaller teams according to the three different<br />

structures: 1. Biopsychosocial, 2. Community- Oriented Primary Care, and 3. Integrated Delivery Systems.<br />

In order to provide a higher quality service, the care station was set up in the community.<br />

275


3.2 System design and implementation<br />

Chih-Yu Lin et al.<br />

In this study, we adopted the prototype model to develop the community health promotion system. Since<br />

system users are mainly healthcare workers, thus they do not have enough knowledge about the system<br />

development. This system includes two subsystems: (1) Community health promotion website (CHPS)<br />

which contains the diseases and health knowledge, Such as diabetes and hypertension. (2) IRS system.<br />

For example, the teachers can use this system to get the feedback from community residents on their<br />

health education progress. The system architecture has shown schematically in figure 1. The system is<br />

developed by using the PHP programming language and MYSQL database.<br />

Figure 1: The system architecture<br />

3.3 Survey instrument<br />

In order to evaluate the system performance, the system was evaluated in two parts: one is on the website<br />

effectiveness, and the other is on the IRS system. We surveyed the user's satisfaction for website and the<br />

IRS by distributing questionnaires to various members and IRS users.<br />

Both questionnaires were developed based on the literature about system satisfaction measurement and<br />

then sent to expert panel for review before being distributed to various members and IRS users. All items<br />

are structured on a Likert five-point scale ranging from 1 (strongly disagree) to 5 (strongly agree) in both<br />

questionnaires.<br />

There are two parts in the website’s questionnaire. Part one of the questionnaire collects respondents’<br />

demographics such as occupation, gender, age, work experience, and education level. Part two is the<br />

measure of end-user computing satisfaction (EUCS) developed by Doll and Torkzadeh and that is<br />

comprised of 12 questions (Doll and Torkzadeh, 1988).<br />

The questionnaire for IRS satisfaction includes four constructs such as perceived usefulness (PU),<br />

perceived ease of use (PEOU), total satisfaction, and perceived enjoyment. The items of first three<br />

constructs were modified from Wang’s study (Wang, 2003), and the items of perceived enjoyment were<br />

based on Agarwal and Karahanna’s research (Agarwal and Karahanna, 2000).<br />

4. Results<br />

4.1 The results of the CHPS website’s satisfaction<br />

The majority of users for CHPS website were hospital employees who join our health promotion activities.<br />

These users included physicians, nurses, other medical staff, and administrative personnel.<br />

276


Chih-Yu Lin et al.<br />

A total of 165 questionnaires were mailed out and 141 returned. The majority of respondents were female<br />

(66.0%), aged over 30 (67.4%), nurses (39.0%), and work experience less than ten years’ (77.3%). The<br />

details of demographics about the respondents are summarized in Table 1.<br />

Table 1: The details of demographics information about the respondents<br />

Item Category (Percentage)<br />

Male (24.1%)<br />

Gender<br />

Female (66.0%)<br />

Missing value (9.9%)<br />

21-30 years old (32.6%)<br />

Age<br />

31-40 years old (44.7%)<br />

More than 40 years (22.7%)<br />

Less than 1 year (9.2%)<br />

1-10 years inclusive (68.1%)<br />

Work Experience<br />

More than 10 years (20.6%)<br />

Missing value (2.1%)<br />

Physician (7.1%)<br />

Nurse (39.0%)<br />

Occupation<br />

Other Medical staff (28.4)<br />

Administrative personnel (25.5)<br />

Senior high school (7.8%)<br />

College (84.4%)<br />

Education<br />

Graduate school (7.1%)<br />

Missing value (0.7%)<br />

We use the EUCS instrument to evaluate the user's satisfaction for the website. The EUCS instrument<br />

comprises five constructs for measuring end-user computing satisfaction: content, accuracy, format, ease<br />

of use, and timeliness. The average score of this questionnaire is 4.3, and all of the items in the<br />

questionnaire are higher than 4.18. That means most people were satisfied to the website. The results<br />

showed that PEOU was the dimension satisfied the most people. It means the operation of the website is<br />

easy for users. The reliability of the instrument was measured by Cronbach’s alpha. The α values is 0.901<br />

which means this questionnaire has high internal consistency. The results of descriptive statistics about<br />

the user's satisfaction for website are summarized in Table 2.<br />

Table 2: The results of descriptive statistics for the user's satisfaction for website<br />

Construct Item Mean S.D. AVE.<br />

Content<br />

Accuracy<br />

Format<br />

Ease of<br />

Use<br />

Timeless<br />

Does the website provide the precise information you need?<br />

Does the information content meet your needs?<br />

Does the website provide reports that conforms to exactly what you need<br />

?<br />

Does the website provide sufficient information?<br />

4.24 0.608<br />

4.18 0.628<br />

4.23 0.569<br />

4.26 0.630<br />

Is the website accurate?<br />

Are you satisfied with the accuracy of the website?<br />

4.29<br />

4.25<br />

0.606<br />

0.638<br />

4.27<br />

Do you think the output is presented in a useful format?<br />

Is the information clear?<br />

4.26<br />

4.31<br />

0.659<br />

0.623<br />

4.29<br />

Is the website user friendly?<br />

4.36 0.668<br />

Is the website easy to use?<br />

4.44 0.669<br />

4.40<br />

Do you get the information you need in time?<br />

4.33 0.615<br />

Does the website provide up-to-date information?<br />

4.27 0.685<br />

4.30<br />

Total 4.30<br />

277<br />

4.23


Chih-Yu Lin et al.<br />

The user's satisfaction for website among groups had significant difference. According to the analysis of<br />

LSD post hoc test, most of the medical professional groups (physicians, nurses, and other medical staff)<br />

have high satisfaction than administration personnel in 4 constructs (accuracy, format, ease of use and<br />

timeless), especially in accuracy. The result implies that the website is more suitable to the medical<br />

professionals because they have more medical knowledge than administration personnel to estimate.<br />

Users with different work experience also had different satisfaction to the website. Group of “< 1 years”<br />

working experience is more satisfied to the website than user who have “1-10 years” or “> 10 years”<br />

working experience in content, accuracy, format, and timeless. The results may be reasonable because<br />

people with more working experience may have higher competence to know what they really need. The<br />

details of different satisfaction among groups are summarized in table 3<br />

Table 3: The details of different satisfaction between groups<br />

Content<br />

Accuracy<br />

Format<br />

Ease of Use<br />

Timeless<br />

Item<br />

*P < 0.05<br />

Occupation Gender Age<br />

1. Physician<br />

2. Nurse<br />

3. M.T.<br />

4. Executives<br />

1. Female<br />

2. Male<br />

1. < 40 years<br />

2. 31-40 years<br />

3. > 40 years<br />

Work<br />

Experience<br />

1. < 1 years<br />

2. 1-10 years<br />

3. > 10 年 years<br />

Education<br />

1. Senior high<br />

school<br />

2. College<br />

3. Graduate school<br />

F value Test F value Test F value Test F value Test F value Test<br />

Q1 0.923 — 8.089 — 0.728 — 1.852 — 0.977 —<br />

Q2 2.621 — 0.284 — 2.101 — 1.485 — 0.118 —<br />

Q3 2.084 — 4.506 — 1.462 — 3.369*<br />

1>2<br />

1>3<br />

0.887 —<br />

Q4 1.286 — 0.134 — 2.773 — 2.122 — 0.597 —<br />

Q5 4.461*<br />

2>4<br />

3>4<br />

0.101 — 3.420* 2>3 3.430*<br />

1>2<br />

1>3<br />

0.181 —<br />

Q6 3.119* 3>4 1.743 — 0.212 — 3.095* 1>3 1.129 —<br />

Q7 2.864*<br />

2>4<br />

3>4<br />

2.853 — 0.335 — 3.666* 1>2 0.869 —<br />

Q8 1.459 — 3.280 — 4.176*<br />

2>1<br />

2>3<br />

2.207 — 0.244 —<br />

Q9 4.376*<br />

Q10 4.767*<br />

Q11 3.049*<br />

1>2<br />

1>4<br />

3>4<br />

1>4<br />

2>4<br />

3>4<br />

1>4<br />

3>4<br />

3.669 — 0.015 — 1.021 — 0.605 —<br />

1.021 — 1.651 — 1.182 — 1.158 —<br />

5.397 — 1.130 — 0.922 — 0.470 —<br />

Q12 1.176 — 0.044 — 0.211 — 4.355*<br />

278<br />

1>2<br />

1>3<br />

0.362 —


Chih-Yu Lin et al.<br />

4.2 The results of the CHPS for IRS satisfaction<br />

The majority of users for IRS are community residents who came to care station to participate the health<br />

promotion activities, there are 27 subjects in our study.The instrument comprised four constructs for<br />

measuring the user's satisfaction for IRS: perceived usefulness, perceived ease of use, perceived<br />

enjoyment, and total satisfaction. The average score of this questionnaire was 4.74, and all items were not<br />

less than 4.22 score, which means most residents are satisfied to use the IRS in the health education<br />

program. The results of descriptive statistics for user's satisfaction of IRS are summarized in Table 4.<br />

Table 4: The results of descriptive statistics for IRS satisfaction.<br />

Construct Item Mean S.D. AVE.<br />

Perceived<br />

usefulness<br />

Perceived<br />

ease of use<br />

Perceived<br />

enjoyment<br />

Total<br />

satisfaction<br />

5. Limitation<br />

The IRS system makes it easy for you to discuss questions with your<br />

teachers.<br />

The content provided by the IRS system is easy to understand.<br />

It is easier to concentrate on the content when the IRS system is used.<br />

4.93 0.385<br />

5.0 0.0<br />

4.93 0.267<br />

The feedback from the IRS system allows the teacher to understand<br />

how well I have educated myself about the diseases.<br />

4.85 0.456<br />

I would like to request the teachers to use the IRS system as frequently<br />

as possible during the classes.<br />

If a teacher uses the IRS system during classes, I will continue to use<br />

4.52 0.509<br />

the IRS system. 4.52 0.509<br />

The remote control function of IRS system is easy to use. 4.41 0.501<br />

The operation of the IRS system is stable. 4.22 0.506<br />

Using the IRS system during classes makes the class more interesting.<br />

4.89 0.320<br />

The IRS system creates a more interactive environment in the class. 4.89 0.320<br />

4.79<br />

4.32<br />

4.89<br />

As a whole, you are satisfied with the IRS system. 4.96 0.192 4.96<br />

Total 4.74<br />

Due to the restriction of resources and time limit, we are currently using it in two different communities. By<br />

using the system we would like educate the communities on the healthy living and how to prevent<br />

ourselves from various diseases. As we are currently using them in two communities, therefore we are<br />

unable generalize the feedback of these two communities to represent the feedback of a larger population.<br />

Therefore, we are not able to identify the performance of the system on a larger scale.<br />

Most of the subjects are currently from community residents who are 60 years old and above and have a<br />

low literate rate. Therefore, it must be spent more time consuming to educate the older community<br />

residents to use our system. If a health organization would like to follow our method, we would like to<br />

recommend them to use them on people who have at least elementary education so the effectiveness of<br />

community residents’ self-health management will reduce the time consumed to understand how to use<br />

our system.<br />

6. Conclusion<br />

The purpose of IRS that applied on the community health promotion and health education is to build a<br />

friendly learning environment to improve community residents’ cognition of health promotion. Through a<br />

teaching tool of IRS, the departments related to community health promotion can be integrated<br />

horizontally and they can support the community health promotion programs continuously. In addition,<br />

building a effective management system of community residents’ health according to the learning<br />

effectiveness and developing a health promotion education program and a community website which are<br />

proper to community characteristic are in order to promote the health care of community residents.<br />

279


Chih-Yu Lin et al.<br />

The results of this study could be a reference for healthcare institutions. It can improve effectiveness for<br />

healthcare institutions to adopt IT to community health promotion activities. Therefore, healthcare<br />

institutions will involve in community health promotion activities actively.<br />

Applying IT into community health promotion activities properly can help healthcare institutions to manage<br />

self-care quality of community residents effectively. Therefore, it might reduce the incidence of chronic<br />

diseases and the medical expenditure on chronic diseases and long-term care.<br />

Combining IT in community health promotion activities in this study could be a reference model. We<br />

suggest that developing a better community health management information system and technology can<br />

improve the health of community residents continuously and conform to the spirit of sustainable<br />

development of community health promotion. In further, it also can fulfill the health policies of localized<br />

medical care, localized health service, localized aging, and preventive medicine.<br />

References<br />

Agarwal, R. and Karahanna, E. (2000) Time Flies When You’re Having Fun: Cognitive Absorption and Beliefs about<br />

Information Technology Usage. Management Information Systems Quarterly, Vol 24, No. 4, pp 665-694.<br />

Doll, W.J. and TorKzadeh, G. (1988) The Measure of End-User Computing Satisfaction. Management Information<br />

Systems Quarterly, Vol 12, No. 2, pp 259-274.<br />

Dorjsuren, B. (2008) Financing health promotion in Japan and Mongolia. Bulletin of the World Health Organization,<br />

Vol 86, No. 11, pp 896-897.<br />

Hancock, T. and Duhl, L. (1986) Promoting Health in the Urban Context.Copenhagen, WHO Regional Office for<br />

Europe (Healthy Cities Paper No.1).<br />

Heilman, J.M., Kemmann, E., Bonert, M., Chatterjee, A., Ragar, B., Beards, G.M., Iberri, D.J., Harvey, M., Thomas, B.,<br />

Stomp, W., Martone, M.F., Lodge, D.J., Vondracek, A., de Wolff, J.F., Liber, C., Grover, S.C., Vickers, T.J.,<br />

Meskó, B. and Laurent, M.R. (2011). Wikipedia: A Key Tool for Global Public Health Promotion, Journal of<br />

Medical Internet Research, Vol 13, No. 1, e14.<br />

Pagliari, C., Sloan, D., Gregor, P., Sullivan, F., Detmer, D., Kahan, J.P., Oortwijn, W. and MacGillivray, S. (2005) What<br />

is eHealth (4): a scoping exercise to map the field. Journal of Medical Internet Research, Vol 7, A9.<br />

Squyres, W.D. (1985) Patient Education and Health Promotion in Medical Care.Palo Alto, CA: Mayfield Publishing Co.<br />

Ottawa Charter (1986), WHO.<br />

Wang, Y.S. (2003) Assessment of Learner Satisfaction with Asynchronous Electronic Learning Systems. Information<br />

& Management, Vol 41, pp 75-86.<br />

Wantland, D.J., Portillo, C.J., Holzemer, W.L., Slaughter, R. and McGhee, E.M. (2004) The effectiveness of<br />

Web-based vs non–Web-based interventions: a meta-analysis of behavioral change outcomes. Journal of<br />

Medical Internet Research, Vol 6, No. 4, e40.<br />

WHO(1994). Action for health in cities. Copenhagen,WHO Regional Office for Europe.<br />

WHO(1978). Alma Ata Declartion.Geneva.<br />

Yajima, S.H., Takano, T. Nakamura, K. and Watanabe, M. (2001) Effectiveness of a community leaders’ programme<br />

to promote healthy lifestyles in Tokyo, Japan. Health Promot International, Vol 16, pp 235-243.<br />

280


Information Security Practices in Zambian Copper Mines: An<br />

Investigation Into the State-of-Practice of Information<br />

Security Within Zambian Copper Mines Based on the ISO/IEC<br />

27002 Standard<br />

Chishala Lukwesa and Christopher Upfold<br />

Rhodes University, Grahamstown, South Africa<br />

g09l2525@campus.ru.ac.za<br />

c.upfold@ru.ac.za<br />

Abstract: Information has become an important aspect of daily operations in the mining industry. This is due to the<br />

mining industry’s increased dependence on information for strategic operations and competitive advantage.<br />

Therefore, the need for reliable and accurate information to achieve this, demands enhanced information security<br />

practices in these organisations. Sound information security practices require an effective information security<br />

program backed by information security policies and frameworks. The research aims to investigate the state-ofpractice<br />

of information security in Zambian copper mines using the ISO/IEC 27002 standard as the foundation of<br />

information security, and to explore leadership perceptions regarding information security practices within these<br />

mines. Available literature on information security practices within the mining industry is used to identify barriers to<br />

effective implementation of information security practices and a conceptual framework is proposed to this effect. This<br />

framework is then verified by responses obtained by means of a survey administered to senior management and<br />

heads of Information Technology in Zambian copper mines, and an in-depth case study of information security<br />

practices in one of the said Zambian copper mines using an ISO/IEC 27002 audit tool. Identified barriers to effective<br />

information security implementation include, differing perceptions between senior management and middle<br />

management, incomplete information asset inventory, poor staff awareness, inadequate information security policies<br />

and procedures, inadequate risk management processes, and inadequate control system policies and procedures. A<br />

framework for enhanced implementation of information security in Zambian copper mines is further proposed. This<br />

framework is made up of proposed considerations and enhancements that can be used to address the barriers<br />

identified in the literature review, based on results from the leadership survey and the ISO/IEC 27002 audit<br />

instrument. This framework, based on ISO/IEC 27002 guidelines and controls, includes full management<br />

commitment to information security, adoption of an information security standard or framework, improved personnel<br />

awareness, complete and up-to-date asset inventory, defined information security governance framework, and<br />

enhanced policies and procedures for industrial control systems.<br />

Keywords: information security, leadership perceptions, state-of-practice, copper mines, ISO/IEC 27002<br />

1. Introduction<br />

The mining industry is a continuous operational and strategic industry, which contributes to both local and<br />

international economies (Teseleanu et al., 2006). This industry, like other manufacturing industries, is<br />

data intensive as it deals with information about processes, product designs, shipping, inventory and<br />

customers (Paladion, 2008). Although sometimes considered to be low-tech with little innovation and<br />

appreciation for information technologies, the industry is best understood as a scale-intensive innovative<br />

industry with its own characteristics (Upstill and Hall, 2007; Pavitt, 1984). Although argued to be slowchanging<br />

and somewhat traditional, the face of the industry has been seen to change. Part of this change<br />

has been in the timing of mining information, which is becoming indispensable to sound policy making,<br />

strategy and forecasts for future changes (Sawada, 2004). The quality of information, its production, use,<br />

flow, accessibility and credibility have been identified as key to developing trust and cooperation amongst<br />

stakeholders in the mining industry. This is coupled with openness and greater transparency in<br />

information production and dissemination throughout the mineral life cycle for enhanced decision making.<br />

(MMSD, 2002: 296). Therefore, just as quality management standards such as ISO 9000 have become a<br />

requirement for doing business in industries such as the mines, internationally recognised information<br />

security standards continue to be accepted and will eventually become a necessity for many mines<br />

(Ernst&Young, 2008). One of these information security standards is ISO/IEC 27002. This standard is<br />

universally accepted and used globally. ISO/IEC 27002 has over the years become the de facto standard<br />

for high level definition of an information security management system as it defines a comprehensive set<br />

of controls that help organisations implement a good information security program (Wessels et al., 2007:<br />

253). ISO/IEC 27002 was used in this research as it provided a framework that was suitable for Zambian<br />

copper mining organisations. ISO/IEC 27002 encompasses 11 information security domains, which are<br />

universally accepted information security domains. These domains were all applicable to current<br />

information security practices in the Zambian copper mining industry.<br />

281


2. Literature review<br />

2.1 Risk management<br />

Chishala Lukwesa and Christopher Upfold<br />

Risk management in the mining industry has been concentrated on safety, environmental, occupational<br />

health, and quality management. With the advent of the information age, the risk management process<br />

has been extended to include physical IT assets. Information that is stored, transmitted and processed<br />

within the same IT infrastructure has, however, not yet been well covered (Pironti, 2008). In view of this,<br />

although the use of information is critical to effective risk management in the mining industry, information<br />

risk management itself has not been one of the high priority risk areas. An integrated or enterprise risk<br />

management approach, which aims to identify, evaluate and manage major corporate risks in an<br />

integrated framework, is, however, now being adopted by the mining industry. This is in order to<br />

encompass all risk areas and not just focus on safety, environmental, occupational health and quality<br />

areas (Fagg, 2009; Olson and Wu, 2008: 3).<br />

2.2 Contingency planning<br />

Organisations are built on similar principles but each has unique characteristics in the way business is<br />

conducted and products and services provided. There are certain components that particular<br />

organisations cannot function without. These components are a combination of people, processes and<br />

technology. There is at least one person, certain technology or information specific business processes<br />

and some critical infrastructure components that are critical to business operations. If their continuity is<br />

not catered for in times of manmade or natural disasters, the business might fail. Therefore, an<br />

organisation should identify the critical components that need to be protected (Seldon, 2009).<br />

Contingency planning in the mining industry covers a lot of areas and issues and may involve production,<br />

distribution, Information Technology (IT) and other functional areas (Musson, 2007: 185). All functional<br />

areas and not just IT should be catered for in the business continuity plan as they need to co-exist at any<br />

given time. Recovery in mining industries also presents challenges, which may not be found in other<br />

industries. These range from government regulations to certification and re-certification requirements<br />

such as ISO 9000 (Musson, 2007: 185; Warner, 2007). It is, therefore, vital for the mining industry to<br />

deploy a business continuity process that encompasses all areas of the enterprise including supplier and<br />

customer networks. Disaster Recovery Plans (DRPs) and/or Business Continuity Plans (BCPs), should,<br />

provide a solution that will account for and ensure the recovery of all business processes cater for an<br />

alternate workspace environment for employees, facilitate ongoing operations within a consistent<br />

timeframe and cover all regulatory requirements (Warner, 2007). Warner (2007) argues that in many<br />

cases, the amount of investment in a DRP or BCP could literally mean the difference between the<br />

business and its shareholders sinking or swimming. A Meta Group report in 2000 revealed that metal and<br />

natural resource industries can lose an average of $580,588 per hour in downtime costs. It is not only<br />

money that can be lost but also reputation, current and potential customers, and customer confidence<br />

(Hinton and Clements, 2002).<br />

2.3 Information Security Governance<br />

Peltier et al. (2005: 240) define Information Security Governance (ISG) as the management structure,<br />

organisation, responsibility, and reporting processes surrounding a successful information security<br />

program. According to Von Solms and von Solms (2009: 24) information security governance is also the<br />

system by which the confidentiality, integrity and availability of an organisation’s electronic assets are<br />

maintained. Mears and von Solms (2004) and Krehnke (2007: 41-43) define the factors that make up the<br />

pillars of information security governance. They all agree that these factors include accountability and<br />

responsibility, ethics, resource allocation and management, security awareness and education,<br />

information security policies, good practice standards, risk management, compliance with legal<br />

requirements, and information sharing.<br />

Employees are a critical factor in information security as they are the first line of defence in any operation<br />

within an organisation. Security Education, Training, and Awareness (SETA) is, therefore, an important<br />

aspect of any information security program (Whitman and Mattord, 2009: 206).<br />

The Zambian government has enacted a number of laws, policies and acts, which are meant to provide<br />

for information security and to which mining organisations and its employees like every other entity in the<br />

country should adhere to. These include the National ICT Policy, Electronic Communications and<br />

282


Chishala Lukwesa and Christopher Upfold<br />

Transactions Act, 2009, and the Copyright and Performance Rights Act No. 44 of 1994 (Ministry of<br />

Communications and Transport, 2006; National Assembly of Zambia, 2009; GRZ, 1994; Campbell, 2008:<br />

260).<br />

Employees in the mining industry also make use of mobile devices such as two-way radios (EPA, 2007),<br />

iPhones and Blackberry handsets and laptops, which create a security loophole in the organisation.<br />

Security strategies, employee training and awareness, and a revised IT infrastructure are required to help<br />

achieve effective mobile security governance (Zhang, 2009).<br />

Failure in corporate governance threatens the future of every organisation (Arjoon, 2005). Information<br />

security should, therefore, be an integral part of corporate governance, aligned with IT governance and<br />

integrated into strategy, concept, design, implementation and operation (ITGI, 2006: 15).<br />

2.4 Information Security standards<br />

Implementation of best practice standards is important in order to ensure that an organisation’s<br />

information security governance practices instil confidence in the stakeholders, shareholders and<br />

customers (Brotby, 2009: 84). Adoption of an information security standard complements an<br />

organisation’s information security policy and assists in making the implementation of the policy more<br />

effective (Peltier, 2002: 70).Just as quality management standards such as ISO 9000 have become a<br />

requirement for doing business in industries such as the mines, internationally recognised information<br />

security standards will continue to be accepted and eventually become a necessity for many mines<br />

(Ernst&Young, 2008). An organisation wanting to implement security controls which are in compliance<br />

with a standard or set of standards requires support from top management and other employees in the<br />

development and implementation process. These standardised policies and guidelines must be<br />

applicable to the organisation’s culture, business and operational practices (HKSAR, 2008).<br />

3. Information Security concerns<br />

Although Zambian copper mining organisations have implemented information security measures in their<br />

organisations, there are certain aspects of information security practices that are considered to require<br />

addressing. A summary of the concerns that were identified in the literature review as barriers to effective<br />

implementation of information security in the Zambian copper mining industry is outlined in table 1 below.<br />

Table 1: Summary of identified information security concerns in the Zambian copper mining industry<br />

Concern Author(s)<br />

1. Lack of or inadequate security policies Bradley, 2008; Welander, 2007<br />

2. Lack of or inadequate security procedures Bajpai and Gupta, 2005; Welander, 2007<br />

3. Poor staff awareness Bajpai and Gupta, 2005; DTT, 2008<br />

4. Inadequate risk management process Pironti, 2008<br />

5. Increased regulation PricewaterhouseCoopers, 2009; WIPO,<br />

2009<br />

6. Dispersed data and information Copans, 2007; PricewaterhouseCoopers,<br />

2008<br />

7. Email threats such as viruses and spam DTT, 2008<br />

8. Inadequate control system policies and procedures Lowe, et al., 2007; Welander, 2007<br />

9. Cultural clashes between IT staff Bartels, 2005<br />

10. Poor asset identification and inventory PricewaterhouseCoopers, 2008<br />

11. Lack of formal mobile security governance Zhang, 2009<br />

12. Irregular or no patching of systems Bartels, 2005<br />

13. Inadequate access control procedures Bajpai and Gupta, 2005<br />

14. Inadequate email policies Bradley, 2008<br />

15. Informal change management procedures Bartels, 2005<br />

16. Informal BCPs DTT, 2008<br />

17. Lack of or irregular testing of contingency plans DTT, 2008; ICMM, 2009<br />

18. Ill-defined crisis communication procedures ICMM, 2009<br />

19. Lack of formal information security governance frameworks DTT, 2008<br />

20. Inadequate integration of information security into<br />

DTT, 2008<br />

employment policies and practices<br />

21. Poor incident response management Helberger, 2009<br />

22. Unsecure application software Tomlinson, 2007: 12; Dobelis, 2007: 46<br />

23. Ill-defined user access privileges Van Holsbeck and Johnson, 2004;<br />

Kairab, 2004: 7<br />

283


Chishala Lukwesa and Christopher Upfold<br />

Concern Author(s)<br />

24. IT-based BCPs Warner, 2007<br />

25. Ill-maintained legacy systems Zoufaly, 2009; Welander, 2009; Shelton,<br />

2009<br />

26. Human error such as social engineering Tomlinson, 2007: 12; DTT, 2008; Piccoli,<br />

2008: 431<br />

27. Incorrect dissemination of information MMSD, 2002<br />

28. Environmental threats to hardware Falco et al., 2002<br />

These concerns were mapped against the eleven (11) ISO/IEC 27007 information security domains and<br />

controls to determine to what extent the ISO/IEC 27002 security domains cater for all or part of the<br />

security concerns.<br />

4. Research methodology<br />

The research was conducted by means of an instrumental case study and includes such data collection<br />

methods as surveys in the form of questionnaires, interviews and document reviews. According to Stake<br />

(1995: 3), an instrumental case study is used when there is a need for a general understanding and there<br />

is a possibility of getting insight into the question by studying a particular case. The case study method<br />

was chosen as it is an intensive study that sheds light on a larger class of classes and uses different<br />

methods to collect various kinds of information and make observations (Gerring, 2007:20; Hamel et al.,<br />

1993: 45). It is through these methods that the object of study is understood (Hamel et al., 1993: 45).<br />

Case study methods also answer the “how” and “why” of research questions and also focus on<br />

contemporary events (Yin, 2003: 5).<br />

Descriptive statistics were used for quantitative analysis of the data. Fisher and Marshall (2009) define<br />

descriptive statistics as “the numerical procedures or graphical techniques used to organise and describe<br />

the characteristics or factors of a given sample”. Two survey instruments were administered. One set of<br />

questionnaires made up of information security concerns identified in the literature review was<br />

administered to senior executives and heads of Information Technology departments in these mines.<br />

These questionnaires were meant to explore leadership perceptions surrounding information security<br />

practices in Zambian copper mines and were administered in four of these mines. An in-depth case study<br />

was then administered to personnel in charge of infrastructure in these mines by means of a<br />

commercially available survey instrument in the form of an ISO/IEC 27002 audit tool procured from<br />

Praxiom Limited. This representation of middle management included human resource and administrative<br />

personnel, and personnel in charge of the various sections of the mining organisation’s Information<br />

Technology department. The audit tool was administered in person by the researcher. Ethical, reliability,<br />

and validity issues surrounding the collection and analysis of data were addressed.<br />

5. Findings and recommendations<br />

Research findings and recommendations that were deemed to be most significant are discussed below<br />

and outlined according to the 11 ISO/IEC 27002 domains. These domains were categorised into<br />

administrative, technical and physical domains. Recommendations were made based on the responses<br />

from the two sets of questionnaires.<br />

Administrative domains include:<br />

5.1 Information Security policy<br />

Senior management believe information security policies in their organisations adequately address<br />

information security requirements, business and organisational requirements and that they also cater for<br />

changing business needs. Middle management, on the other hand felt that the information security policy<br />

was in need of attention. 75% of senior management believe information security policies have been<br />

implemented in their organisations while 25% do not. All these respondents further believe their<br />

organisation’s information security policies meet their information security requirements, business<br />

objectives, and business requirements. Middle management on the other hand revealed that the<br />

information security policy is not kept up to date.<br />

Recommendations<br />

Information security policies provide support for information security governance and serve as a<br />

backbone for information security programs in an organisation. Thus they should be regularly reviewed at<br />

284


Chishala Lukwesa and Christopher Upfold<br />

planned intervals or when a significant information security change is made in the organisation in order to<br />

ensure that they continue to reflect the organisation’s current business and information security<br />

requirements. Users should also be notified of subsequent updates to information security policies and<br />

procedures.<br />

5.2 Organising Information security<br />

Both senior management and middle management believe only part of organisational data has been<br />

consolidated. Therefore, not all data and information in these organisations can be accounted for. 25% of<br />

senior management firmly believe that all organisational data has been consolidated and accounted for in<br />

their organisation while 50% could neither agree nor disagree as to whether this was the case in their<br />

organisations. An additional 25% on the other hand, does not believe that all their organisation’s data and<br />

information have been adequately consolidated for it to be appropriately accounted for.<br />

Recommendations<br />

Information assets should be clearly identified, defined, and responsibility allocated for each information<br />

asset. This includes definition of responsibilities for specific information security processes such as<br />

business continuity. Allocation of responsibilities should be done according to the information security<br />

policy and further documented.<br />

5.3 Asset management<br />

Both senior and middle management believe asset management still requires attention.<br />

Recommendations<br />

All organisational information assets should be identified and inventoried. This inventory should be<br />

regularly updated and should include among other things, BCPs, contracts and agreements, archived<br />

information, research information, training materials and user manuals, and audit trails.<br />

5.4 Human resources management<br />

Senior management believe information security is adequately integrated into employment policies and<br />

procedures while middle management on the other hand felt that the organisation’s security roles and<br />

responsibilities are not documented or communicated in such areas as employees’ job descriptions.<br />

However, both senior and middle management agree that personnel do not receive adequate information<br />

security education, training, and awareness.<br />

Recommendations<br />

Effective management of personnel information security roles and responsibilities ensures that their use<br />

of information processing facilities is appropriately managed and minimises insider threats. Therefore,<br />

management should undertake the responsibility of ensuring that all personnel receive the appropriate<br />

level of awareness, training, and education, which is relevant to their job role, responsibility and skills.<br />

Job descriptions should also be used to document and address employees’ information security roles and<br />

responsibilities.<br />

5.5 Business continuity management<br />

Senior management endorses business continuity strategies in these organisations and certainly the<br />

impacts that failures could have on the organisations’ processes have been evaluated. However, the<br />

inclusion of information security requirements in the BCPs and regular reviews of these BCPs still require<br />

attention. Both senior and middle management are aware that some aspects of business continuity<br />

management still require addressing. 50% of senior management firmly believe that contingency plans to<br />

deal with system failures in their organisations have been established and are regularly tested and<br />

reviewed. 25% could neither agree nor disagree as to whether this is the case, and a further 25% do not<br />

believe these contingency plans have been established or are regularly reviewed and tested. Additionally,<br />

50% of senior management believe their organisations have detailed BCPs in place which specify actions<br />

that should be taken in the event of a disaster. A further 25% could neither agree nor disagree as to<br />

whether this was the case and 25% do not believe these detailed BCPs are in place in their<br />

organisations.<br />

285


Recommendations<br />

Chishala Lukwesa and Christopher Upfold<br />

Effective business continuity management reduces the chances of interruptions to business activities and<br />

minimises the impact these interruptions would have on the business. A single framework for BCPs<br />

should be adopted and maintained to ensure that all plans are consistent and that they consistently<br />

address information security requirements, along with the identification of priorities for testing and<br />

maintenance. These plans should also be regularly tested, reviewed, and updated to ensure that they are<br />

realistic and can be used in a real-life scenario.<br />

5.6 Compliance<br />

Senior management believes their organisations adhere to regulatory requirements while middle<br />

management on the other hand, felt that requirements for intellectual property rights are in need of<br />

attention. Middle management felt that procedures that deal with intellectual property rights and<br />

proprietary software products as well as user awareness regarding the protection of these rights still<br />

require attention.<br />

Recommendations<br />

Compliance with legal, statutory, contractual, and regulatory requirements is necessary for the<br />

organisation to avoid legislative implications. All regulatory, statutory, and contractual requirements that<br />

are applicable to the organisation should be explicitly identified, documented and updated regularly. This<br />

should also be done for each information system. Appropriate procedures to ensure compliance with<br />

legislative, contractual, and regulatory requirements regarding intellectual property rights and proprietary<br />

software products should also be implemented.<br />

5.7 Information Security incident management<br />

Both senior and middle management believe the establishment and implementation of incident response<br />

plans still requires attention. 50% of senior management firmly believe that their information security<br />

incident reporting responsibilities and channels have been defined while the other 50% could neither<br />

agree nor disagree as to whether these incident reporting responsibilities and channels have been<br />

defined in their organisations.<br />

Recommendations<br />

Information security incidents should be incorporated into personnel awareness programs as examples of<br />

possible incidents, and how they could be handled and prevented in the future. Additionally, procedures<br />

that would be used to handle different information security incidents should be established. These<br />

procedures should cater for, among others, malicious code, denial of service attacks, and breaches of<br />

integrity.<br />

Technical domains include:<br />

5.8 Access control<br />

Both senior and middle management believe mobile security governance requires attention. 50% of<br />

senior management believe mobile security governance measures have been put in place to address<br />

mobile computing and communications in their organisations while 25% could neither agree nor disagree<br />

as to whether this was the case in their organisations. A further 25% do not believe these security<br />

measures have been put in place.<br />

Recommendations<br />

The mobile computing and communication policy should be used to address all mobile devices in use in<br />

the organisation coupled with personnel training to raise awareness on the risks associated with mobile<br />

computing. This policy should include requirements for physical protection, access control, backup<br />

facilities, cryptographic controls, and virus protection.<br />

286


Chishala Lukwesa and Christopher Upfold<br />

5.9 Information Systems acquisition, development and maintenance<br />

Both senior and middle management believe emphasis on information security requirements during the<br />

acquisition, development, and implementation of information systems still requires attention. 50% of<br />

senior management believe that information security requirements are emphasised during the<br />

procurement and development of application software in their organisations. 25% could neither agree nor<br />

disagree as to whether this is so as there have been cases of violations in the procurement process. A<br />

further 25% believe these information security requirements are not emphasised in their organisations.<br />

Recommendations<br />

Controls to be used for ensuring authenticity and message integrity in applications should be identified<br />

and implemented. This includes implementation of cryptographic controls and development of a policy on<br />

the use of these controls. Cryptographic techniques such as digital signatures should be used to achieve<br />

confidentiality, authenticity and integrity, and non-repudiation of information.<br />

5.10 Communications and operations management<br />

Senior management believe adequate electronic communication policies and procedures have been<br />

developed in their organisations while middle management on the other hand, felt that electronic<br />

communication procedures and controls still require attention.<br />

Recommendations<br />

Formal policies, procedures, and controls should be used to address the control of the exchange of<br />

information through all communication facilities. These controls should include:<br />

Use of cryptographic controls in information exchange to ensure the authenticity, confidentiality, and<br />

integrity of information.<br />

Restrictions on the forwarding of electronic mails to external mail destinations.<br />

Physical domains include:<br />

5.11 Physical and environmental security<br />

Adequate protection of an organisation’s assets from physical and environmental threats helps prevent<br />

loss, misuse, theft or compromise to an organisation’s activities. While senior management believe their<br />

organisations’ equipment is protected from environmental threats, middle management on the other hand<br />

felt that although this was the case, security of equipment from physical threats still requires attention.<br />

Recommendations<br />

Physical access controls should be implemented to only allow authourised access to secure areas. This<br />

should include regular review and update of access rights to secure areas. All personnel and visitors<br />

should be required to wear visible identification and personnel should notify security personnel when they<br />

encounter anyone not wearing visible identification. Measures for the protection of information processing<br />

equipment used offsite should be enhanced and should include:<br />

Disguising portable computers when travelling.<br />

Using risk assessments to identify and further implement home working controls including clear desk<br />

policies.<br />

5.12 Summary of findings<br />

Findings from senior management revealed that information security concerns deemed to be most<br />

prevalent in Zambian copper mines include:<br />

Poor staff awareness.<br />

Inadequate control system policies and procedures.<br />

Findings from the in-depth analysis of information security practices in a Zambian copper mine revealed<br />

the following information security concerns:<br />

Lack of or inadequate information security procedures<br />

287


Poor staff awareness<br />

Poor asset identification and inventory<br />

Chishala Lukwesa and Christopher Upfold<br />

Lack of formal mobile security governance frameworks<br />

IT-based Business Continuity Plans<br />

Inadequate email policies<br />

Informal BCPs<br />

Ill-maintained legacy systems<br />

These findings revealed that in most cases, senior management perceptions do differ from those of<br />

middle management. This implies that senior management may not be fully committed to the information<br />

security process within these organisations or may not be fully aware of the extent to which their<br />

organisations meet International best practice standards such as those outlined in the ISO/IEC 27002<br />

standard.<br />

6. Conclusion<br />

Zambian copper mining organisations face numerous challenges when it comes to safeguarding their<br />

information assets. These organisations have evolved from a background where business efficiency<br />

mostly affected the physical component of business operations to one where newer technologies have<br />

led to an increased dependence on information for efficient and strategic operations This evolution<br />

requires that management take up the ultimate responsibility of driving the efficient use of this information<br />

which includes effectively safeguarding information assets in order for this efficiency to be achieved.<br />

Information security programs, therefore, have to be effectively implemented and should emphasise<br />

allocation of information security responsibilities, implementation of Security Education, Training and<br />

Awareness (SETA), and adoption of an information security framework or standard. Adoption of a<br />

standard or framework will ensure that information security practices in these organisations are<br />

standardised and evidence from their information systems can be admissible in court in case of legal<br />

implications. Although the security of information can never be foolproof as the nature and use of<br />

information systems continues to evolve and new risks are identified, measures can still be put in place to<br />

reduce the effect these risks will have on business activities and the organisation’s operations.<br />

7. Directions for future research<br />

There is a need to conduct further in-depth case studies by administering the ISO/IEC 27002 Audit Tool<br />

Questionnaires in more than one copper mining organisation in Zambia. This would most likely yield<br />

further in-depth understanding of information security practices in these organisations, especially from a<br />

middle management perspective.<br />

A possible limitation of this work could relate to the ‘halo’ effect in which perhaps a slightly optimistic view<br />

of questionnaire responses could have been given by respondents. In addition, administering a similar<br />

audit tool to the users themselves would give a more complete picture of information security practices in<br />

these organisations from a senior management, middle management and user perspective.<br />

References<br />

Arjoon, S. (2005) “Corporate Governance: An Ethical Perspective”, Journal of Business Ethics, Vol 61, pp 343-352.<br />

Bajpai, S. and Gupta, J. P. (2005) “Site Security for Process Industries”, Journal of Loss Prevention in the Process<br />

Industrie,. Vol 18, No. 4-6, pp 301-309.<br />

Bartels, N. (2005) “Securing the Process Industries”, [Online], http://www.mbtmag.com/article/193165-<br />

Securing_the_process_industries.php.<br />

Bradley, C. (2008) “Top Five Steps to Proactively Manage Email Risk and Keep Compliance in Check”, [online],<br />

http://www.itsecurityjournal.com/index.php/Latest/Top-Five-Steps-to-Proactively-Manage-Email-Risk-and-Keep-<br />

Compliance-in-Check.html.<br />

Brotby, K. (2009) Information Security Governance, John Wiley and Sons, Hoboken.<br />

Campbell, C. (2008) Legal Aspects of Doing Business in Africa, Yorkhill Law Publishing, Salzburg.<br />

Copans, G. (2007) “ERP Integrator Company Provides Consolidation, Governance to Mine Operations”, [Online],<br />

http://www.miningweekly.com/article/erp-integrator-company-provides-consolidation-governance-to-mineoperations-2007-10-05.<br />

Deloitte Touche Tohmatsu (DTT) (2008) “Gaining momentum. The 2008 Energy & Resources Global Security<br />

Survey”, [online], http://www.deloitte.com/view/en_AU/au/industries/energyresources/energy/article/9f6de23a4b101210VgnVCM100000ba42f00aRCRD.htm.<br />

288


Chishala Lukwesa and Christopher Upfold<br />

Dobelis, E. (2007) Expert Systems for Business Decision on Security Requirements. In: Meersman, R., Tari, Z.,<br />

Herrero, P., Herrero, H. M. (eds). On the Move to Meaningful Internet Systems 2007: OTM 2007 Workshops,<br />

Springer, Berlin.<br />

EPA (2007) “Two-Way Radios”, [online],<br />

http://cfpub.epa.gov/safewater/watersecurity/guide/productguide.cfm?page=twowayradios<br />

Ernst & Young (2008) “Ernst & Young‘s 2008 Global Information Security Survey: Moving Beyond Compliance”,<br />

[online], http://www.ey.com/Global/assets.nsf/Russia_E/GISS_2008_EN/$file/GISS_2008_EN.pdf.<br />

Fagg, S. (2009) “Mining, Financial Services Leading Risk Charge”, [online]<br />

http://www.riskmanagementmagazine.com.au/articles/fb/0c02c6fb.asp.<br />

Falco, J., Stouffer, K., Wavering, A. and Proctor, F. (2002) “IT Security for Industrial Control Systems”, [online],<br />

http://www.isd.mel.nist.gov/documents/falco/ITSecurityProcess.pdf.<br />

Fisher, M. J. And Marshall, A. P. (2009) Understanding Descriptive Statistics. Australian Critical Care. 22: 93—97<br />

Gerring, J. (2007) Case Study Research: Principles and Practices. New York: Cambridge University Press<br />

Government of the Republic of Zambia (GRZ) (1994) “Copyright and Performance Rights Act, 1994”, [online],<br />

http://portal.unesco.org/culture/en/files/30413/11425068613zm_copyright_1994_en.pdf/zm_copyright_1994_en.<br />

pdf.<br />

Hamel, J., Dufour, S. And Fortin D. (1993) Case Study Methods. Newbury Park: SAGE Publications<br />

Helberger, C. (2009) “10 Key Information Security and Compliance Activities for 2010”, [online],<br />

http://www.itbusinessedge.com/cm/community/features/guestopinions/blog/10-key-information-security-andcompliance-activities-for-2010/?cs=34705.<br />

Hinton, W. and Clements, R. (2002) “Are You Managing the Risks of Downtime?”, Disaster Recovery Journal. Vol<br />

15, pp 3. [online], http://www.drj.com/articles/sum02/1503-12.html.<br />

Hong Kong Special Administrative Region (HKSAR) (2008) “An Overview of Information Security Standards”,<br />

[Online], http://www.infosec.gov.hk/english/technical/files/overview.pdf.<br />

International Council on Mining and Metals (ICMM) (2009) “Case Study: Communication Plans for Crisis<br />

Management”, [online], http://www.icmm.com/page/722/communication-plans-for-crisis-management.<br />

IT Governance Institute (2006) Information Security Governance: Guidance for Boards of Directors and Executive<br />

Management (2e), ITGI, Rolling Meadows<br />

Kairab, S. (2004) A Practical Guide to Security Assessments, Auerbach Publications, Boca Raton.<br />

Krehnke, D. C. (2007) Corporate Governance. In: Tipton, H. F. and Krause, M. (eds). Information Security<br />

Management Handbook (6e), Taylor and Francis group, Boca Raton.<br />

Lowe, J., Lasky, J., Gilbert, D. and Sheil, K. (2007) “Protecting Industrial Process Control, Automation, and SCADA<br />

Systems From Cyber Threats”, SPE Middle East Oil and Gas Show and <strong>Conference</strong>, [online],<br />

http://www.onepetro.org/mslib/servlet/onepetropreview?id=SPE-105696-MS&soc=SPE.<br />

Mears, L. And von Solms, R. (2004) “Corporate Information Security Governance: A Holistic Approach”, [online],<br />

http://icsa.cs.up.ac.za/issa/2004/Proceedings/Research/041.pdf<br />

Mining, Minerals, and Sustainable Development Project (MMSD) (2002) Breaking New Ground: Mining, Minerals,<br />

and Sustainable Development: The Report of the MMSD Project, IIED, London.<br />

Ministry of Communication and Transport (2006) “Zambia National Information and Communication Technology<br />

Policy”, [online], http://www.mct.gov.zm/pdf/ict.pdf.<br />

Musson, M. (2007) Business Continuity Strategies for Manufacturing and Logistics. In: Hiles, A. (ed). The Definitive<br />

Handbook of Business Continuity Management, John Wiley & Sons: Chichester.<br />

National Assembly of Zambia (2009) “The Zambian Parliament”, [online],<br />

http://www.parliament.gov.zm/index.php?option=com_docman&task=cat_view&gid=115&Itemid=113.<br />

Olson, D. L. and Wu, D. D. (2008) Enterprise Risk Management, World Scientific, Toh Tuck Link, Paladion<br />

(2008)“Manufacturing”, [online], http://www.paladion.net/Manufacturing.html.<br />

Pavitt, K. (1984) Sectoral Patterns of Technical Change: Towards a Taxonomy and a Theory. Research Policy. Vol<br />

13, pp 343-373.<br />

Peltier, T., Peltier, J. and Blackley, J.(2005) Information Security Fundamentals, CRC Press, Boca Raton.<br />

Peltier, T.R. (2002) Information Security, Policies, Procedures, and Standards, CRC Press, Boca Raton.<br />

Piccoli, G. (2008) Information Systems for Managers: Texts and Cases, John Wiley & Sons Inc, Hoboken.<br />

Pironti, J. P. (2008) “Securing Information Infrastructure:Expert Advice on Evaluating the New Risks and Structuring<br />

Your Defenses”, [online], http://www.iparchitects.com/wp-content/uploads/Securing-Information-<br />

Infrastructure.pdf.<br />

Pricewaterhouse Coopers (2008) “The Global State of Information Security® 2008: Energy and Mining Security<br />

Chiefs Have Advanced in Many—but not all—Critical Security and Privacy Arenas”, [online],<br />

http://www.pwc.com/extweb/insights.nsf/docid/A8659F991CA54234852574DB005CFA9E/$File/Global_Info_Sur<br />

vey_Energy.pdf.<br />

Sawada, K. (2004) Mining Information and Mining Policy. Journal of the Mining and Materials Processing Institute of<br />

Japan: Vol 120, No. 9, pp 532-534.<br />

Seldon, A. (2009) “You Cannot Use it if it is Gone. Hi-Tech Security Solutions: The Industry Journal for Security &<br />

Business Professionals”, [online], http://securitysa.com/article.aspx?pklArticleId=5621&pklCategoryId=106.<br />

Shelton, R. (2009) “An Integrated Approach to Legacy IT Modernization”’ [online],<br />

http://www.acs-inc.com/assets/0/24/818/824/eb1056b2-9058-4170-82cf-696257083213.pdf.<br />

Stake, R. E. (1995) The Art of Case Study Research. Thousand Oaks: Sage Publications.<br />

Teseleanu, G., Maghear, T., and Andras, A. (2006) “The Need for Innovation and Technology Transfer in Mineral<br />

Industry”, [online], http://opi4.opi.org.pl/repository/41fa6dc088dd71307f6cbcd93a3ca1b6Ot9EO9.pdf.<br />

289


Chishala Lukwesa and Christopher Upfold<br />

Tomlinson, R. F. (2007) Thinking About GIS: Geographic Information System Planning For Managers (3e), ESRI, Inc<br />

, Redlands.<br />

Upstill, G. and Hall, P. (2007) Innovation in the minerals industry: Australia in a global context. Resources Policy, Vol<br />

31, No. 3, pp 137-145.<br />

Van Holsbeck, M. and Johnson, J. Z. (2004) “Security in an ERP World”, [online], http://www.netsecurity.org/article.php?id=691.<br />

Von Solms, R. and Von Solms, S. H. (2009) Information Security Governance, Springer, New York.<br />

Warner, D. (2007) “Disaster Recovery Solutions for Business Continuity within the Manufacturing Industry”, [online],<br />

http://www.plasticsbusinessmag.com/wp/?p=35.<br />

Welander, P. (2007) “10 Control System Security Threats”, [online], http://www.controleng.com/article/269570-<br />

10_Control_System_Security_Threats.php.<br />

Welander, P. (2009) “Securing Legacy Control Systems”, [Online],<br />

http://www.controleng.com/index.php?id=483&cHash=081010&tx_ttnews[tt_news]=866.<br />

Wessels, P. L., Grobbelaar, E., Mc Gee, A. & Prinsloo, G.T.M. (2007) Information Systems in a Business<br />

Environment (4e), Lexis Nexis, Durban.<br />

Whitman, M. E. and Mattord, H. J. (2009) Principles of Information Security (3e), Course Technology, Boston.<br />

World Intellectual Property Organisation (WIPO) (2009) “What is WIPO?”, [online], http://www.wipo.int/aboutwipo/en/what_is_wipo.html.<br />

YIN, R. (2003) Case Study Research: Design and Methods. Thousand Oaks: SAGE<br />

Zhang, R. (2009) “Steps for Achieving Proper Mobile Security Governance”, [online],<br />

http://www.computerworld.com/s/article/9133117/Steps_for_Achieving_Proper_Mobile_Security_Governance?t<br />

axonomyId=&intsrc=kc_feat&taxonomyName=.<br />

Zoufaly, F. (2009) “Issues and Challenges Facing Legacy Systems”, [online],<br />

http://www.developer.com/mgmt/article.php/11085_1492531_1.<br />

290


Proposing an Optimized Change Management Process by<br />

Analyzing ITSM Frameworks<br />

Negar Madani, Ali Suzangar, Mohamad Kajbaf, Shirin Nasher and Mehrdad<br />

Kalantarian<br />

Infoamn, Tehran, Iran<br />

n.madani@infoamn.com<br />

a.suzangar@infoamn.com<br />

m.kajbaf@infoamn.com<br />

sh.nasher@infoamn.com<br />

m.kalantarian@infoamn.com<br />

Abstract: This paper proposes an optimized change management process for simplified organizational transition by<br />

evaluating and analyzing widely accepted ITSM frameworks such as ITIL, MOF v3, MOF v4 & FITS from different<br />

aspects including but not limited to the benefits and deficiencies of work flow, roles and responsibilities, key<br />

performance indicators & ARCI matrix. IT service management (ITSM) as a sub-discipline of IT management plays<br />

an important role in achieving the organizations goals and objectives. Many frameworks have focused on ITSM, and<br />

as a best practice Information Technology Infrastructure Library (ITIL) is the most widely accepted approach to ITSM<br />

and it is owned by United Kingdoms’ Office of Government Commerce (OGC). ITIL v3 is presented in five main<br />

categories and each category contains its related processes. Based on ITIL, Microsoft Corporation developed a<br />

framework for IT service management named Microsoft Operational Framework (MOF). Although MOF 3.0 is mostly<br />

in accordance with ITIL framework, MOF 4.0 is presented by three main areas and a management section that<br />

covers the activities in each area. MOF 4.0 focus is on insuring the alignment of business and IT throughout the<br />

organizations activities. One of the most important aspects in ITSM frameworks is the manner in which changes are<br />

managed in an organization. This article includes evaluating and analyzing the change management process in the<br />

above mentioned frameworks from different aspects. As identifying the role of individuals and groups is the most<br />

critical aspect of any process, the first step is to assign roles and responsibilities. This is done by using an ARCI<br />

matrix which presents the individuals and groups responsible, accountable, consulted and informed. In order to<br />

identify the deficiencies of the ARCI matrix for each framework, the matrixes are analyzed and compared and a new<br />

ARCI matrix is introduced. The next step is to define the sequence in which the actions are taken, mentioned as a<br />

work flow. The work flow of each framework is analyzed and the importance and disadvantages of each action is<br />

discussed from a practical point of view. After defining the change process, each framework applies key performance<br />

indicators for assessing its performance. These indicators are listed and compared and the most effective and<br />

efficient indicators are determined. This evaluation and the lessons learned, resulted in an optimized change<br />

management process flow with a particular ARCI matrix and revised key performance indicators which covers the<br />

identified gaps and deficiencies and can be used to benefit organizations in adopting changes and making successful<br />

transitions.<br />

Keywords: IT management, ITIL, change management, ITSM framework<br />

1. Methodology<br />

This article focuses on the comparison of change management processes in ITIL, MOF 3, MOF 4 and<br />

FITS.<br />

At first change management is introduced briefly, and the reasons for its importance and its goals and<br />

objectives have been presented in the second section.<br />

The Third section includes a table in which change management processes of ITIL, MOF3, MOF4 and<br />

FITS are presented. The processes are compared based on the main activities required for a change<br />

management process, and then the benefits, advantages and disadvantages of each process are<br />

discussed.<br />

The disadvantages of the change management processes are solved and an optimized change<br />

management process is presented in section 4, this optimized process has the advantages of all of the<br />

discussed change management processes.<br />

One of the important categories in section 4 is defining responsibility types for the optimized change<br />

management process known as RACI matrix, and assigning it to specific roles. Due to the fact that there<br />

has not been a RACI matrix in the studied processes, it is not gained by reviewing the change<br />

291


Negar Madani et al.<br />

management processes, but it is the result of a practical experience of implementing change<br />

management process.<br />

There are many other requirements for an optimized change management process, which mostly depend<br />

on the specification of each organization, therefore they are mentioned briefly as the additional<br />

requirements of a change management process. Outstanding results that have been achieved by<br />

optimizing the change management process are explained in the conclusion.<br />

2. Introduction to change management<br />

Change management plays a key role in any organization. A change may be triggered from inside or<br />

outside of an organization. The manner in which the change is managed and implemented is the subject<br />

of change management process (Duck, J. D. 2000). It is not always easy to make a change due to<br />

various reasons, for instance a single organization may have varied entities, and changing an entity may<br />

be reflected in other entities. The role of the staff is also very important for the change process to be<br />

smooth. Change management requires a broad set of skills including analytical, financial, business,<br />

system and communication skills. To make a change many different types of models, methods,<br />

techniques and tools are used, however all of the different methods follow a certain structured organized<br />

process which has proven to be successful for leaving the current state and moving to the final state<br />

(OGC 2007). The change management would help an organization and its processes to be stable.<br />

Change management is responsible for the changes made to the IT infrastructure and services, and<br />

ensures that changes are made with minimal risk or consciously taken risk. The main benefits of following<br />

standard processes or proper change management is reduction of the number of potential incidents and<br />

problems associated with each change, simple return to a stable configuration, reduction of back-outs<br />

needed, etc. Like any other process, the main elements of Change management process are as<br />

following:<br />

Process flow, which defines the steps and sequence of activities required to successfully deploy a<br />

change.<br />

Process owner, which is the person in full control of change management process, gathers process<br />

performance data and is consulted in activities related to the process.<br />

RACI matrix, which determines the responsibility types assigned to different roles, and<br />

Key performance indicators (KPI) are the most important metrics which indicate and help authorities to<br />

manage the efficiency and effectiveness of the processes.<br />

2.1 Scope<br />

There are many different approaches to the scope of change, yet changes are divided into two main<br />

categories, reactive changes and proactive changes (OGC 2007).<br />

Reactive changes are changes made in order to align the organization with the business and technology<br />

requirements (OGC 2007).<br />

Proactive changes are changes which are made in response to every day requirements like errors (OGC<br />

2007).<br />

The scope of change management is based on the specifications of each organization. In most<br />

organizations the scope of changes are defined based on the scope of configuration items (CI). If any CI<br />

is altered due to an action, the action is considered as change therefore, it should be done through<br />

change management process (Haber, L., 2003).<br />

3. Process flow<br />

In change management an important step is to define a standard process which presents the activities<br />

required to successfully deploy a change. Different processes have been presented for change<br />

management. The change management processes of ITIL, MOF4, MOF3 and FITS are compared in the<br />

table below. The first column presents the activities mentioned in ITIL V3 service transition publication.<br />

The next two columns present Microsoft Operations frameworks (MOF). Because of the different<br />

292


Negar Madani et al.<br />

approach to MOF4, there are still reasons to consider MOF3 as well. And the last column presents the<br />

British Educational Communications and Technology Agency point of view on a change management<br />

process.<br />

Table 1: Comparison between change management processes<br />

ITIL MOF4 MOF3 FITS<br />

1-Create and record the RFC<br />

(Recommended to use a tool that<br />

automatically stores the initial<br />

data)<br />

2- Review RFC and change<br />

proposal:<br />

3-Assess and evaluate the<br />

change<br />

4- Authorize the change<br />

5- Coordinate change<br />

implementation<br />

6- Review and close change<br />

1: Baseline the<br />

configuration.<br />

2: Initiate the change.<br />

3: Classify the change.<br />

4: Approve the change.<br />

5: Develop and test the<br />

change.<br />

6: Release the change.<br />

7: Validate and review<br />

the change.(update<br />

database)<br />

No steps No steps<br />

1: Identify change<br />

1-Change initiator<br />

2-Change request 2: Create request for<br />

3-Change<br />

change<br />

classification<br />

4-Change<br />

3: Approval to proceed<br />

authorization<br />

4: Plan and prepare change<br />

5-Change<br />

5: Peer review<br />

development<br />

6: Approval to implement<br />

6-Release<br />

management<br />

7-Change review<br />

7: Implement<br />

NO steps defined for<br />

validation<br />

8: Update configurationmanagement<br />

database<br />

No steps defined closure 9: Closure<br />

As presented in the above table, the firs activity in change management is base-lining and recording the<br />

current state, before taking any actions. Base-lining is clearly cited in MOF4 (1.Baseline the<br />

configuration) and is implicitly mentioned in ITIL (1- Create and record the RFC).<br />

The next activity which is alike in all processes, is initiating a change by a Request for Change (RFC). A<br />

RFC is a formal proposal for a change to be made and includes details of the proposed change, and may<br />

be recorded on paper or electronically (British Educational Communications and Technology Agency<br />

2004). In most of the above processes when the first step is determined as “initiating a change”, it is<br />

assumed that base-lining is an obvious fact and will be carried out before making any changes to the<br />

environment. The person which initiates the change is called the change initiator and is informed of the<br />

progress and the result of the change. Therefore in this activity the need for a change is recognized and<br />

is requested by a change initiator.<br />

The next step in MOF3 and MOF 4 is classifying the change, but in ITIL the change is reviewed first and<br />

then it is classified in another step. Classifying the change in MOF 4 includes identifying the priority and<br />

the category of the change (Microsoft 2008).<br />

The priority includes:<br />

Low: The change can wait until the appropriate time.<br />

Medium: Because of the impact on the change, the change should be attended as soon as possible<br />

High: The change should be made as soon as it has been tested.<br />

Emergency: The change should be made immediately; therefore many of the approval and the<br />

development steps are abbreviated.<br />

The category includes:<br />

Standard change: This change has minimal business impact and has a known set of release<br />

processes that has been proven to be successful.<br />

Minor change: This change has a low business impact and affects a small percentage of users and<br />

resources.<br />

Significant change: This change has a moderate impact on the business. It might involve downtime of<br />

services.<br />

293


Negar Madani et al.<br />

Major change: This change has the greatest impact on business, with a high risk and high cost.<br />

Emergency change: This category is high risk because of the urgency of release and the minimal<br />

time in which to test it.<br />

Unauthorized change: This level involves changes that occur outside of the agreed-to change<br />

management policies or that are specifically forbidden.<br />

MOF 3 has a similar category with a slight difference. In MOF 3 changes with the emergency priority, are<br />

assumed emergency changes. No category is defined for unauthorized changes in MOF 3 change<br />

management process (Microsoft 2004).<br />

As mentioned above, ITIL change management process has addressed classifying the RFC in step 4<br />

“Authorize the change”. ITIL has the same priority levels and category; moreover it contains a risk<br />

categorization matrix based on impact and probability (OGC 2007), as follows:<br />

Risk category 1: high impact/high probability<br />

Risk category 2: high impact / low probability<br />

Risk category 3: low impact / high probability<br />

Risk category 4: low impact / low probability<br />

The next step is to review the request and to approve it. This activity is indicated by different titles in each<br />

process, however the concept and explanations for each title declares the meaning of “reviewing and<br />

approving the RFC”.<br />

After a change has been correctly prioritized and categorized by the change manager, depending upon<br />

the category and priority, the change is routed to the appropriate authority for approval. A significant or<br />

major change is usually approved by presenting the change to a team of reviewers known as the change<br />

advisory boards (CAB) (Microsoft 2004). The CAB is composed of many different people with many<br />

different skills for the many different features of the organization. One member of the CAB will be chosen<br />

as the change owner which will be held accountable for the results of the change.<br />

It is the responsibility of the CAB to determine if the change should be approved and scheduled, be<br />

refused and ended or be returned for further clarification. Emergency changes are normally reviewed by<br />

an emergency committee of the CAB typically known as the Emergency change advisory board (ECAB).<br />

Standard changes are approved automatically and progress directly to the change development and<br />

release phase. Minor changes can be approved by the change manager without reference to the CAB.<br />

The next step of the change process is the preparation of the change. ITIL addresses preparing the<br />

change with the title of “Assess and evaluate the change” which includes: The seven Rs of change<br />

management, risk categorization, allocation of priority, evaluation of change, change planning and<br />

scheduling and assessing remediation (OGC 2007).<br />

The stated activities are mostly covered in MOF 4, MOF 3 and FITS with slight differences. For instance<br />

MOF 4 and MOF 3 address the allocation of priority in step three. In step four, MOF 3 schedules the<br />

change, appoints an owner, prepares and develops the change (Microsoft 2004), while MOF 4 designs<br />

the change, identifies configuration dependencies, builds and tests the change (Microsoft 2008).<br />

FITS on the other hand attends to the above mentioned activities in step four, five and six of the change<br />

management process: 4.Plan and prepare change, 5.Peer review, 6.Approval to implement. “Plan and<br />

prepare change” includes: details of change, impact of change, risk of change, remediation plan, change<br />

schedule, authorities names (British Educational Communications and Technology Agency 2004).<br />

The next step of the change process is making the actual change. This is usually done by passing the<br />

RFC to the relevant technical group. Therefore this step is largely a coordination role in change<br />

management.<br />

ITIL has named this step “coordinate change implementation”. MOF 3 and FITS simply forward the RFC<br />

to release management process which is responsible for releasing changes. From then on the role of<br />

change management is to observe the progress of releasing the change.<br />

294


Negar Madani et al.<br />

MOF 4 does not have an independent process for release management. In MOF 4 the release process is<br />

a combination of Deploy SMF which is a sub-process of Delivery phase, and Change and configuration<br />

SMF which is a sub-process of Manage phase. The change is applied through the steps defined in the<br />

sub-processes.<br />

The last step is to review and close the change. On completion of the change, the results should be<br />

documented and reported to appropriate authorities. To confirm that the change has met its objectives, a<br />

post implementation review (PIR) should be carried out. Depending on different types of changes, after a<br />

certain time has elapsed, changes should be reviewed (OGC 2007).<br />

ITIL addresses reviewing and closing the change in a single step with fewer details (OGC 2007), while<br />

closing the change is mentioned implicitly in MOF 4 (Microsoft 2008).<br />

In MOF 3, reviewing the change is determined in an explicit process with detailed activities and<br />

explanations for each types of change (Microsoft 2004).<br />

4. Optimized change management process<br />

The reason for presenting an optimized change management process is to solve the problems identified<br />

in the above processes and to present a process which has the advantages of all of the mentioned<br />

processes. The advantages and disadvantages of the above processes have been identified through<br />

reviewing the requirements of a successful change management process presented in ISO publications<br />

(ISO 2005), management controls (Institute of Internal Auditors 2005) and ITSM success measures<br />

(Haber, L., 2003).<br />

The optimized change management process includes:<br />

Process flow: This process flow is obtained through considering the main requirements of a<br />

successful change and, the comparison of different change management processes.<br />

RACI Matrix: The concept of the RACI matrix for change management does not exist in any of the<br />

above processes; therefore the RACI matrix has been the result of implementing change<br />

management.<br />

Key performance indicators: The KPIs of all other processes have been compared and revised in<br />

order to present a new list, which contains the appropriate KPIs for the optimized change<br />

management process.<br />

Additional requirements: Different types of organizations require different types of templates and<br />

instructions for a successful change management process; therefore it is not possible to pre define all<br />

the additional requirements. However, a brief list of templates and instructions which were required<br />

for the implementation of change management has been presented.<br />

What is unique about this optimized change management process: The outstanding achievements<br />

that have made this process unique and what distinguishes it from the other processes has been<br />

presented and thoroughly explained.<br />

4.1 Process flow<br />

In order to present an optimized change management process, main activities that are required for a<br />

successful change (ISO 2005) and the sequence in which the activities should be presented (Duck, J. D.<br />

2000) has been identified.<br />

The activities are selected based on the advantages of all the reviewed processes. And the order of the<br />

activities is arranged in a way, which will require minimal duplication.<br />

The steps and activities of the optimized change management process are selected based on the above<br />

analysis and the consideration of different environments and organizations; therefore, the steps and<br />

activities are arranged in a way to address the identified deficiencies mentioned in the above analysis.<br />

The table below presents the optimized change management process flow, which the first column<br />

indicates the main steps required to perform a change, while the column on the right indicates the<br />

activities required to fulfill each step.<br />

295


Table 2: An optimized change management process<br />

Negar Madani et al.<br />

Step Activity<br />

1-1) Create and record the RFC<br />

1-2) Review RFC<br />

1) Record and review the RFC<br />

1-3) Approval to proceed<br />

1-4) Baseline and identify CIs effected<br />

1-5) Assess and evaluate the change<br />

2) Change classification<br />

2-1) Prioritize the change<br />

2-2) Categorize based on risk<br />

2-3) Classify the change<br />

3) Authorize change 3-1) Select the authorized person/advisory board<br />

4) Change schedule and escalation<br />

4-1) Schedule change<br />

4-2) Appoint a responsible person<br />

5) Release management 5-1) Release management<br />

6) Review and close change<br />

4.2 RACI matrix<br />

6-1) Select the review team<br />

6-2) Schedule the review<br />

6-3) Review<br />

6-4) Check for Standard change<br />

6-5) Close the Change<br />

In order to determine the required roles and responsibilities for the optimized change management<br />

process, the determined roles and responsibilities in each framework has been reviewed and after<br />

considering the different environments, organizations and their requirements and also taking the<br />

optimized change management processes steps and activities into account, the bellow RACI matrix is<br />

presented.<br />

The RACI matrix has not been included in any of the mentioned change management processes;<br />

therefore it was not possible to present a comparison. The presented matrix is the result of implementing<br />

the optimized change management process.<br />

An RACI Matrix categorizes tasks into four responsibility types, which each type is assigned to specific<br />

roles per process step and activity.<br />

Responsible: This type indicates the person who actually carries out the task.<br />

Accountable: This type is usually assigned to one person through a process and that person is<br />

accountable to the correct and through completion of the task.<br />

Consulted: This type is assigned to authorities who are consulted and asked for advice in specific<br />

sections of the process.<br />

Informed: This type indicates the authorities who should be informed of the progress of a change.<br />

As determined in the table, the change initiator is accountable and responsible for requesting a change in<br />

the format of an RFC. The change manager is held accountable for the rest of the process. The first two<br />

steps are also, mostly the responsibility of the change manager or another person assigned by him. But<br />

in the rest of the process the change manager is just accountable while the responsibilities are assigned<br />

to the CAB. When the change is ready for deployment, the incident and problem management should be<br />

informed and ready to confronting any problems occurred due to the change. And from then on, it is the<br />

responsibility of the release management to deploy the change. After the change has been deployed, it<br />

has to be reviewed and checked, and the change initiator should be informed of the results.<br />

296


Table 3: Change management RACI matrix<br />

Change management Activities<br />

Negar Madani et al.<br />

1-1) Create and record the RFC A/R<br />

1-2) Review RFC A/R C<br />

1-3) Approval to proceed A/R C/I<br />

1-4) Baseline and identify CIs effected A R<br />

1-5) Assess and evaluate the change A/R<br />

2-1) Prioritize the change A/R C<br />

2-2) Categorize based on risk A/R<br />

2-3) Classify the change A/R<br />

3-1) Select the authorized person/advisory board A/R I<br />

4-1) Schedule change A R C/I<br />

4-2) Appoint a responsible person A R<br />

5-1) Release management A I R<br />

6-1) Select the review team A R<br />

6-2) Schedule the review A R<br />

6-3) Review A I R C<br />

6-4) Check for Standard change A R<br />

6-5) Close the change A I I R I I<br />

4.3 Key performance indicators<br />

The metrics determined as key performance indicators, should have specific meanings and should help<br />

with identifying trends (Haber, L., 2003). It is important that the measures indicate the rate of achieving<br />

business goals (Institute of Internal Auditors 2005). The KPIs presented in the reviewed processes have<br />

been compared. And lots of similarities have been identified between the KPI lists. The appropriate key<br />

performance indicators for the optimized change management are:<br />

Increase in the number of successful changes<br />

Reduction of the cost of changes<br />

Reduction in the number of unauthorized changes<br />

Reduction in the number of emergency requests<br />

Reduction in problems and incidents related to a change<br />

Reduction of astonished personnel due to uninformed changes<br />

Reduction in the number of unnecessary changes<br />

Increase of identified standard changes<br />

Reduction of the time of attending to a change request<br />

Reduction of the time of implementing a change<br />

4.4 Additional requirements<br />

The key elements of change management have been presented in the above sections; however several<br />

templates and documents are required for a successful change management process. The most common<br />

templates which should be prepared before lunching change management process have been cited<br />

below:<br />

A request for change template which includes fields for unique ID, requester details, date/time fields,<br />

change information, escalation and decision results, etc.<br />

A change schedule table containing the date and time of all changes.<br />

Terms and instructions for unauthorized changes<br />

297<br />

Change manager<br />

Change initiator<br />

Configuration<br />

manager<br />

CAB<br />

Incident/Problem<br />

manager<br />

Release<br />

management


Instructions for determining standard changes<br />

Instructions for selecting CAB members<br />

Instructions for the CAB agenda<br />

Negar Madani et al.<br />

The voting logic and the method by which the CAB will approve the change<br />

4.5 What is unique about the optimized change management process<br />

The main achievements and benefits of following this optimized process for implementing changes in an<br />

organization are cited bellow. These statements indicate the strengths of this process and how it adds<br />

value to the organization and the IT infrastructure from different points of view.<br />

1. Improved sequence of steps: in order to avoid unnecessary repetition of activities, the sequences<br />

of the steps are assessed and the sequence with the best results has been selected. As mentioned in<br />

classifying the change in section 2.3, In MOF3 and MOF4 before achieving an official approval for<br />

proceeding, the change is classified, which if the change is not approved it will result in the waste of<br />

resources. Plus there is no certain step for scheduling the change and appointing a responsible<br />

person for each change, which in the optimized change management process is placed in step four.<br />

The last but not the least is the final step of MOF3 and MOF4 which do not close the change<br />

properly. If a change isn’t closed properly the organization will soon face many problems due to the<br />

unfinished changes. Therefore the last step is the step where all the involved people are informed<br />

and the change is properly closed and recorded.<br />

2. Flexible and clear process activities<br />

3. Identifying standard changes: One of the main activities which is not covered in the above<br />

processes is the identification of standard changes. There should be a certain step and certain<br />

instructions for identifying standard changes. In the optimized change management process, before<br />

closing a change, it is checked to see if the change can be categorized as a standard change. This is<br />

the perfect step for identifying standard changes because up to here, the change has been<br />

thoroughly deployed and the specifications of the change are clearly understood; therefore the<br />

decision of categorizing this change as a standard change can be made based on sufficient<br />

information.<br />

4. Base-lining and identifying CIs effected: One of the disadvantages for a change management<br />

process is the lack of base-lining. Usually base-lining is not taken seriously and its importance is not<br />

recognized until a change has failed. As this common mistake is repeated in many organizations,<br />

base-lining is mentioned in the first step, right after the change is approved to proceed.<br />

5. Authorizing the change based on priority, category and classification: Clearly authorizing the<br />

change should be based on the specifications of a change. And because the activities should be<br />

chosen based on the type of the change this is a very important step and should be done carefully.<br />

Classifying the change was thoroughly discussed in section 3.3 and is presented as a major step in<br />

the optimized change management process.<br />

6. Appointing a specific owner responsible for each change: like every other process, a change<br />

management process owner is determined which is in full control of the activities, gathers process<br />

performance data and is consulted in activities related to the process<br />

7. Performing well planned and effective reviews: Reviewing a change is one of the important<br />

activities in any change. A review is consisted of a review team, review schedule and the results of<br />

the review. The optimized change management process has taken care to cover all the required<br />

steps for a successful review, which clearly is lacked in the other processes.<br />

5. Conclusion<br />

Many organizations depend on IT and therefore an appropriate IT service management is essential for<br />

any organization. There are several frameworks for ITSM including ITIL, MOF3 and MOF4, etc., with<br />

each framework focusing on a specific aspect of IT service management.<br />

The mentioned frameworks have been implemented and their advantages and disadvantages have been<br />

discovered in time. These disadvantages include, deploying change through too many steps or<br />

insufficient steps, not addressing certain requirements such as identifying standard changes, the lack of<br />

an accountable person, etc.<br />

298


Negar Madani et al.<br />

Therefore an optimized change management process has been presented in this article, which based on<br />

the detailed explanations in section 4.5, which links the final findings with the problems, will result in more<br />

success, less disruption and loss, increased revenue, reduced costs and achieving the business<br />

objectives. Also this process is designed in a way that requires a minimum amount of time, budget and<br />

effort for implementing a change.<br />

References<br />

British Educational Communications and Technology Agency (2004) Change management: Millburn Hill Road,<br />

Science Park, Coventry CV4 7JJ.<br />

Duck, J. D. (2000) Managing change: the art of balancing, Harvard Business Review, November–December.<br />

Haber, L., (2003) How do you measure ITSM success. Jupitermedia Corporation.<br />

ISO (2005) ISO/IEC 20000-1 Information Technology – Service Management Part 1: Specification,<br />

ISO (2005) ISO/IEC 20000-2 Information Technology – Service Management Part 2: Code of Practice.<br />

Institute of Internal Auditors (2005) Global Technology Audit Guide 2: Change and Patch Management Controls:<br />

Critical for Organizational Success, Institute of Internal Auditors Altamonte Springs, Florida, US.<br />

Microsoft (2008) Microsoft® Operations Framework version 4., Microsoft Publications, USA.<br />

Microsoft (2004) Microsoft® Operations Framework version 3., Microsoft Publications, USA.<br />

OGC, (2007) Service Transition. ITIL Service Management Practices v3., London: The Stationary Office.<br />

OGC, (2007) ITIL Service Management Practices v3., The Stationary Office: London.<br />

299


The use of RFID and Web 2.0 Technologies to Improve<br />

Inventory Management in South African Enterprises<br />

Sizakele Mathaba¹, Nomusa Dlodlo², Quentin Williams³ and Mathew Adigun³<br />

¹ ² ³Council for Scientific and Industrial Research, Meraka Institute, South Africa<br />

¹ ³University of Zululand, South Africa<br />

smathaba@csir.co.za<br />

sizamath@gmail.com<br />

ndlodlo@csir.co.za<br />

qwilliams@csir.co.za<br />

madigun@pan.uzulu.ac.za<br />

profmatthewo@gmail.com<br />

Abstract: Cost-effective inventory management includes balancing the cost of inventory with its profit. Most business<br />

owners fail to recognize the value of the cost of carrying inventory, which include not only the direct cost of storage,<br />

insurance and taxes but also the cost of money tied up in inventory. Running inventory using paper-based systems,<br />

Excel files and traditional enterprise software is a costly and resource-intensive approach that may not even address<br />

the appropriate issues for most businesses. It is with this in mind that this research proposes taking advantage of the<br />

Internet of Things (IoT) and Web 2.0 tools in the management of inventory .IoT promotes the communication of<br />

things/object through sensors. On the other hand Web 2.0 tools promote the communication of people through their<br />

phones or desktop computers. The collaboration of these two technologies could improve inventory management. A<br />

comprehensive literature survey is conducted on inventory management functionalities. IoT and Web 2.0<br />

technologies are then mapped to the identified inventory management functionalities. As a result the research<br />

proposes inventory management architecture. The paper looks at the architecture of a system that fully integrates the<br />

technical advantages of Radio Frequency Identification (RFID) and IoT, in collaboration with web 2.0 tool, twitter, for<br />

loss prevention and as an enabler for locating misplaced stock, anti-counterfeiting of stock, etc. The system will focus<br />

on South African Enterprises as a developing country in Africa.<br />

Keywords: Internet of things (IoT), Radio Frequency Identification (RFID), Web 2.0 tools, inventory management,<br />

South African Enterprises, Twitter<br />

1. Introduction<br />

Supply chain management is the well-organized management of the end-to-end process. It begins with<br />

the design of the product or service and ends with the time when it has been sold, consumed, and in the<br />

end, discarded by the end user. The complete process involves product design, procurement, planning<br />

and forecasting, production, distribution, fulfilment, after-sales support, and end-of-life disposal (Michael<br />

and McCathie, 2005). Inspection and anti-counterfeiting for merchandise are vital contents of supply<br />

chain management, enterprises which need to open the market completely, have to offer consumers a<br />

real-time information inquiry platforms on commodity, and to let consumers have a clear understanding of<br />

the entire life cycle of the goods they consumed and to give them a convenient way to identify the<br />

accuracy of the products. Unfortunately, few enterprises have systems in place that can monitor whole<br />

cycle of products currently. To resolve this difficulty, this paper recommends a products monitoring<br />

information system based on RFID and IoT.<br />

The paper is structured as follows. Section 2 is the literature review on the IoT. Section 3 is the research<br />

methodology, Section 4 represent the rationale of the study, while Section 5 looks Web 2.0. Technology<br />

in general. Section 6 is the proposed architecture of the inventory management system that fully<br />

integrates the technical advantages of RFID and IoT, in collaboration with web 2.0 tools twitter, for loss<br />

prevention and as an enabler for locating misplaced stock, anti-counterfeiting of stock, etc. The system<br />

will focus on South African Enterprises as a developing country in Africa.<br />

2. Background and context of the study<br />

2.1 Internet of Things<br />

The IoT is an incorporated part of the Future Internet and could be defined as a dynamic worldwide<br />

network infrastructure a with self configuring ability based on criterion and interoperable communication<br />

protocols where physical and virtual things have distinctiveness, physical characteristics, and virtual<br />

personalities and use intelligent interface, and are seamlessly integrated into the information network(de<br />

Saint- Exupery, 2009).<br />

300


Sizakele Mathaba et al.<br />

Figure 1 is an example illustrating how IoT works in warehouse management. Traditionally, operations in<br />

warehouse management are processed manually, with information gathered by hands and captured<br />

through keyboards, voice entry or barcodes and integration through human- machine -interface. The<br />

introduction of IoT, involving RFID sensors & actuators, changed the laborious processing of warehouse<br />

operations. Data entry is now automated, which results in accurate information for informed decisions.<br />

Figure 1: IoT in warehouse management (adapted from Fleisch, 2010).<br />

Figure 2 illustrates how the IoT functions.<br />

Figure 2: Illustration of Internet of Things (adapted from Shen, 2010)<br />

It consist of tags which are programmed into an object, each tag contains product information such as the<br />

expiry date, object pressure, temperature, prices etc. Sensors are embedded into these tags and<br />

301


Sizakele Mathaba et al.<br />

readers, and they enable data transfer from the tags to the readers. Readers access data from the tags in<br />

a code form. IoT middleware filters repeating and irrelevant data and sends the code to the local server.<br />

The object information will be available to the internet through the remote server. Object Naming Service<br />

(ONS) works similar to the Domain Name Service (DNS), it points out the servers storing object<br />

information.<br />

In the IoT, objects are autonomous entities in companies, information and social processes where they<br />

will be enabled to interact among themselves. They also interact with the environment by exchanging<br />

data and information sensed on the environment, while reacting autonomously to the real/physical world<br />

events and influencing it by running processes that trigger actions and create services with or without<br />

direct human intervention. The interfaces in the structure of services facilitate communications with these<br />

smart things over the Internet, query and change their state and any information related with them, taking<br />

into account security and privacy matters(de Saint- Exupery, 2009).<br />

Typical Applications of IoT:<br />

“Toyota (South Africa), Carrier tagged to streamline manufacturing and vehicle tracking. The tags are<br />

intended to remain with the vehicle throughout its life cycle” (Baudin, 2005).<br />

A car which notifies you when it must have a service.<br />

A smoke detector which is directly connected to local fire station.<br />

Washing machine twittering when it finishes washing etc. There are quite many areas where IoT can<br />

be applied.<br />

2.2 Radio Frequency Identifiers<br />

RFID tags are divided into two general types, active and passive, depending on their supply of electrical<br />

power. Active RFID tags consist of their own power source, normally an on-board battery. Passive RFID<br />

tags get power from the signal of an external reader. RFID readers also come in active and passive<br />

selections, depending on the kind of tag they read (Intermec Technologies, 2007).<br />

2.2.1 Active tags<br />

Active tags contain their own power source. They send out a stronger signal, and readers can access<br />

them from further away. The on-board power source makes them big and costly, and thus active RFID<br />

systems normally work well on big objects tracked over lengthy distances. Low-power active tags are<br />

generally a little larger than a deck of playing cards. Active tags can stay inactive until they come in range<br />

of a receiver or can continuously transmit a signal. As a result of their on-board power source, they can<br />

function at higher frequencies, normally 455 MHz, 2.45 GHz, or 5.8 GHz, relying on the application’s read<br />

reach and memory needs. Readers can converse with active RFID tags across 20 to 100 meters<br />

(Intermec Technologies, 2007).<br />

2.2.2 Passive tags<br />

On the other hand, passive tags are very economical, they can cost as little as 20 cents a piece, and new<br />

technologies are continuously manufacturing them on a low-price for integration into general materials<br />

and products. Since passive tags are inexpensive, they are expected to be the starting point in the<br />

expansion of RFID implementations especially in South Africa as a developing country. Besides to their<br />

low cost, passive tags can also be reasonably small. Recent antenna technology confines the smallest<br />

functional passive tag simply to the size of a quarter. The larger the tag, the larger the read range. At<br />

present, passive RFID tags contain about 2 Kbits of memory which is quite small to hold much more<br />

complex information than identification and history information. Technology in the nurture of RFID is<br />

continuously improving. Consequently, the amount of information and capabilities of RFID tags will<br />

increase over time, thereby allowing RFID tags to ultimately hold and transmit enough information<br />

(Weinstein, 2005).<br />

Passive-tag readers can continuously transmit its signal or transmit it when required. Once a tag moves<br />

across the reader’s range, it accepts an electromagnetic signal from the reader through the tag’s<br />

antenna. The tag then keeps the energy from the signal in an on-board capacitor. This process is called<br />

inductive coupling. Once the capacitor has made sufficient charge, it can be able to power the RFID tag’s<br />

circuitry, which transmits a modulated signal to the reader. The return signal consists of information<br />

stored in the tag. Low-frequency tags (less than 100 MHz) send information by releasing energy from the<br />

capacitor to the tag coils in altering strengths over time, which affects the radio frequency produced by<br />

302


Sizakele Mathaba et al.<br />

the tag. The reader perceives these varying waves and can use these variances to demodulate the code.<br />

(Weinstein, 2005).<br />

In higher-frequency tags (greater than 100 MHz), the tag transmits the signal by means of backscatter, in<br />

which the tag’s circuit adjust the resistance of the tag’s antenna. This change in resistance makes a<br />

transmission of Radio Frequency (RF) waves, which the reader can accept and demodulate. Passive<br />

tags usually work at frequencies of 128 KHz, 13.6 MHz, 915 MHz, or 2.45 GHz, and have read ranges of<br />

only few inches to 30 feet. Frequency selection depends on the system’s environment, kind of material<br />

the signal have to pass through , and the system’s essential read range. RFID tags can be enclosed in<br />

numerous materials. Plastics are the most familiar material for RFID, making identification cards for<br />

building entrance, credit cards, or bus fares (Weinstein, 2005 and Intermec Technologies, 2007).<br />

Figure 3 represents the RFID components and illustrates how these components works in any given<br />

environment. Tags which usually very small in size, are attached to any given object and they send out<br />

signals which are received by the reader through antenna, which is an enabler for tags and readers to<br />

transmit information. The information is sent to a back-end computer system for processing.<br />

Figure 3: RFID system components (adapted from Yan et.al, 2008)<br />

3. Methodology<br />

A comprehensive literature survey was conducted on inventory management functionalities, and to<br />

identify research work that has been covered in the IoT and web 2.0 in enhancing the management of<br />

inventory in South African Enterprises. IoT and Web 2.0 are then mapped to the identified inventory<br />

management functionalities.<br />

The research question is:<br />

How can IoT and web 2.0 technologies be utilized to improve inventory management systems in South<br />

African enterprises?<br />

Objectives of the study<br />

Review literature on inventory management<br />

Identify systems that have adopted IoT technologies that can be utilised in inventory management.<br />

Come up with inventory management architecture.<br />

4. Rationale of the study<br />

Stock outs result in much loss for enterprises. It is estimated at 30%, which affects retail sales by<br />

between 5 and 18%. In most cases, the stock maybe available but misplaced (Xin, 2009). This is a huge<br />

loss for enterprises. South Africa is not only becoming more aware of RFID, but is also progressively<br />

rising in the uptake of this technology. RFID vendors are becoming specialised in providing specific<br />

applications and catering their offerings to particular industries. Diversity of factors currently influences<br />

South Africa's demand for RFID applications namely:<br />

The need to reduce theft,<br />

303


Fraud and counterfeit products.<br />

Sizakele Mathaba et al.<br />

The RFID technology is able to keep track and trace the location of products precisely than current<br />

technologies in use, which leads to a reduction in losses. The ability of technology to increase efficiency,<br />

government support for RFID projects and international policies that require RFID applications on specific<br />

goods for import, such as beef products, are also influencing the adoption (online, url:<br />

http://dataweek.co.za/article.aspx?pklArticleId=5027&pklCategoryId=31 ).<br />

The slow adoption of RFID in South Africa presents diverse opportunities for retailers. The most vital is<br />

the availability of abundant opportunities that have not yet been exploited but are perfect for early mover<br />

advantage. Mission to roll out RFID applications in different countries are also on the rise. Supply chain,<br />

asset management and mining are amongst the areas expected to increase the demand for RFID<br />

applications, and actions in these areas are on the rise in many African countries. Despite the numerous<br />

opportunities for RFID applications, there are hindering factors to its expansion. The most remarkable is<br />

the high cost of installing the application, particularly for low value volume products. Most enterprises in<br />

South Africa are small to medium enterprises with low quantity of manufacturing. RFID is costly for such<br />

businesses to implement. Consequently, many sectors continue to use alternative solutions such as bar<br />

codes.<br />

The RFID technology is ineffective if function under certain conditions i.e. Transponders do not function<br />

well when tagged on iron objects or wet surfaces. This creates complications for manufacturing and other<br />

industries that use a variety of metallic objects (Dane et. al, 2010). The low level of alliance among<br />

industry participants is also slowing the market's development. Having agreed that RFID requires a<br />

standard of infrastructure, mostly in power and telecommunications, the demand is affected in many<br />

countries and regions facing infrastructural complications. Regardless of various technological<br />

challenges, the cost of RFID tags has started to drop. This movement is expected to carry on as more<br />

advances in the manufacturing of low cost RFID continue. The continued drop in prices will be a key<br />

issue in growing the demand for RFID in South Africa. Low cost tags will be particularly engaging to low<br />

value volume things, where the market for RFID is currently small. This will offer solution for a variety of<br />

enterprises of all sizes. Participants in market have need of working together to resolve interoperability<br />

and other concerns that have an effect on the use of RFID applications, such as discovery of a general<br />

approach to the industry principles(online, url: http:// dataweek.co.za ) .<br />

The forecast for RFID technology in South Africa are not as rosy as those for biometrics. It is doubtful that<br />

RFID in SA will see the same adoption rate as seen in the US, as local retailers are unlikely to set up<br />

RFID go-ahead like they did in the US. On top of that, the charge of importing tags and the difficulty of<br />

deployment are likely to put off many potential RFID customers, says Max Stone, distribution and partner<br />

manager at Motorola Southern Africa's Enterprise Mobility business(online, url:<br />

http://www.autoidlabs.ch/).<br />

RFID in South Africa is well established in some niche environments, such as access control i.e.<br />

Biometrics, tollgates. South African regulator ICASA only permitted suitable EPC (Electronic Product<br />

Code) RFID frequencies for SA last year, so major deployments are yet to happen. RFID has been<br />

deployed very selectively in South Africa; the cost of the tags has been too high for retail organisations to<br />

see advantage from using this type of real-time tracking. However, this may change as time elapses<br />

(online, url: http://www.iweek.co.za/ViewStory.asp?StoryID=201372 ).<br />

5. Web 2.0<br />

Web 2.0 technologies outline the beginning of the subsequent creation of web-based applications. It<br />

allows web applications to be created, that are more operationally rich and quick to respond than the<br />

usual static pages of traditional web technologies. It’s also enabling content to be produced and shared in<br />

real time, with end-users commonly able to add content to applications themselves (O’Reilly, 2007). This<br />

implies that Web 2.0 technologies support open communications and provide users the freedom to share<br />

their suggestions and opinions.Most organizations of all types and sizes and from all industry verticals<br />

have noticed the explosive growth on the web of social and community sites in the consumer space such<br />

as MySpace, YouTube, and the deluge of Web 2.0 sites. Enterprises have observed the move of major<br />

Web players such as Amazon, eBay, Live, Google, and Yahoo to include social and community<br />

elements, and the interest and demand that this has created. Now they are enthusiastically considering<br />

and in several cases constructing portals in communities and businesses for their own organizations.<br />

Web 2.0 is moving to enterprises (Kittowski et.al, 2009).<br />

304


Sizakele Mathaba et al.<br />

Organizations are interested in using Web 2.0 practices mainly in two places, which is within the business<br />

to advance competency and production, and from the organization to the customers to improve revenue<br />

and customer satisfaction. The use of Web 2.0 within organizations is called Enterprise 2.0 and is likely to<br />

be the first area where Web 2.0 will be used by organizations. Enterprises are using Web 2.0<br />

technologies to mainly communicate with customers (advertising), business partners and potential<br />

employees, allowing them to achieve the goal of true real-time collaboration among these parties. This<br />

can increase output and provides enterprises with a way to easily promote their products. The creation of<br />

online communities and blogs or wikis to initiate conversations and share knowledge is proving to be<br />

particularly interesting to enterprises. Enterprises are already using the Web 2.0 technologies inside the<br />

organization, for communication with employees and customers and marketing, etc. IoT technologies, on<br />

the other hand promote the communication among objects without any human intervention leading to a<br />

reduction of labor costs. This study proposes that IoT can be collaborated with web 2.0 technologies for<br />

communication purposes with humans. The objects will communicate with humans for updates through<br />

web 2.0 technology Twitter, to bridge the communication divide between objects and humans.<br />

6. Proposed architecture for inventory management<br />

Enterprise architecture (EA) presents a bright approach to support business strategy execution. EA is an<br />

outline that can direct organizational activities in the technological sense, and supports the business<br />

strategy from the lowest level of process. This outline helps to facilitate an organization’s business vision<br />

by encapsulating multiple views of an enterprise and provides a general understanding of what the<br />

organization aims to achieve. In addition, EA view the enterprise from a holistic and more complete<br />

architectural point of view (Chuang and Loggerrenberg 2010). Figure 4 shows the proposed architecture<br />

for inventory management. It is architecture of a system that fully integrates the technical advantages of<br />

RFID and IoT, in collaboration with the web 2.0 tools twitter, for loss prevention and as an enabler for<br />

locating misplaced stock, anti-counterfeiting of stock, availability of stock on shelves etc.<br />

In this architecture, RFID basically serves as a replacement for the bar code scanners which are normally<br />

used to track products and shipments in similar ways. This architecture fully integrates the technical<br />

advantages of RFID and web 2.0 to provide feedback on the process to end user. Since the application of<br />

FRID require no human intervention, it is essential that the end user is updated in the whole process.<br />

The RFID system consists of three fundamental components. Initially, the RFID tag is attached to an<br />

asset or product in the inventory. The tag contains information about the particular asset or product and<br />

also may include sensors. The next component is the RFID reader, which communicates with the RFID<br />

tags. The last component is the backend system, which links the RFID readers to a centralized<br />

database/server. The centralized database will store information, such as price, for each RFID tagged<br />

item. In this proposed architecture for inventory management, the passive tags will be used due to their<br />

low cost. Among the functionalities expected to be performed by this system includes:<br />

Checking the availability of stock on shelves.<br />

Identifying misplaced stock on shelves.<br />

Identifying expired/ ruined stock.<br />

Identifying counterfeit products.<br />

Sending updates to the relevant user.<br />

Electronic Product Code (EPC) is a unique global identifier of each product in IoT technologies which is<br />

used to track and trace products (Yan, 2008). The EPC RFID readers will be placed among the shelves<br />

and the products will be programmed with EPC RFID enabled tags. EPC RFID tags will send out the<br />

signal which will be received by the EPC RFID readers in the radio frequency field. The readers will<br />

receive the signal through their antennas and transmit the stored information, i.e. validation, tracking,<br />

counts, and error messages to the EPC middleware. The EPC middleware will filter out the repeating and<br />

irrelevant information. Thereafter, information will be sent to the local server. The local server computer<br />

system will twitter the information to the user i.e. reports on inventory, aggregate counts, errors occurred,<br />

misplaced stock etc. through the use of web 2.0 technology. The end user/owner will receive the<br />

notification on inventory through his/her phone. This system gives effective technical reference for<br />

enterprise managers to monitor whole process of inventory without them being physically involved in the<br />

process. The consumers of the products will benefit also in this proposed architecture. They can query<br />

information about the product on the remote server using the EPC (Electronic Product Code), the ONS<br />

305


Sizakele Mathaba et al.<br />

(Object Naming Service) is network system which works similar to the DNS (Domain Name Service).It<br />

spots out servers storing information in the internet.<br />

Figure 4: The proposed architecture for inventory management<br />

7. Conclusion<br />

Internet of Things integrated with the RFID technology and web 2.0 technologies can assist enterprise<br />

owners in managing their inventory, i.e. using them in monitoring the stock validity, stock on shelves,<br />

misplaced stock etc. The use of web 2.0 tools could play a major role in keeping enterprise owners<br />

posted about what is happening on the inventory without them being physically there and helping them to<br />

make informed decisions, and to know urgent matters which may need their attention immediately. Web<br />

2.0 tools bridge that divide of objects and humans. As a result, this study encourages South African<br />

enterprises to actively promote the development procedures of the RFID technology with web 2.0 tools<br />

and the Internet of Things to improve the inventory management in their enterprises. The IoT have the<br />

drawbacks though i.e. the cost of the technology is a major concern for developing countries like South<br />

Africa, but that can addressed by major investments and collaboration with developed countries if<br />

possible.<br />

306


8. Acknowledgments<br />

Sizakele Mathaba et al.<br />

This research was sponsored by Council for Scientific and Industrial Research (CSIR) MERAKA,<br />

Pretoria, South Africa and Internet of Things Capability Group under supervision of Dr Louis Coetzee<br />

(Group Manager).<br />

References<br />

Baudin, M. (2005) Consultant MMTI Manufacturing Management & Technology Institute: RFID Applications in<br />

manufacturingDraft ,Palo Alto, California.<br />

Chuang C, Loggerrenberg J,( 2010), Challenges Facing Enterprise Architects: a South African Perspective,<br />

Department of Informatics, and University of Pretoria, South Africa.<br />

Dane, H. and Michael, K.and Wamba, F. S., (2010) RFID Enabled Inventory Control Optimization: A Proof of<br />

Concept in Small-to-Medium Retailer, University of Wollongong, Australia.<br />

de Saint – Exupery A,(2009) Internet of Things Strategic Roadmap, CERP-IoT EC, Europe.<br />

Frost and Sullivan, (2008) “The potential for RFID applications in Africa” [Online 03 Nov 2010], Dataweek, Electronics<br />

and Communications Technology, http://dataweek.co.za/article.aspx?pklArticleId=5027&pklCategoryId=31<br />

Franz-Kamissoko, L. (2009) “SA lets its fingers do the talking”, [Online 09 Nov 10], The business technology news<br />

magazine for the empowered enterprises, http://www.iweek.co.za/ViewStory.asp?StoryID=201372<br />

Fleisch,E.(2010) “Auto-ID Labs” , [Online 12 Nov 10] University of St Gallen, http://www.autoidlabs.ch/<br />

Kittowski, F. and Klassen, N. and Faust, D. and Einhaus, J (2009) “A comparative Study on the use of Web 2.0 in<br />

Enterprises” Proceedings of I-KNOW ’09 And I-SEMANTICS’09.<br />

Michael, K. McCathie, L. (2005) , The Pros and Cons of RFID in Supply Chain Management, Proceedings of the<br />

International <strong>Conference</strong> on Mobile Business (ICMB’05) IEEE, University of Wollongong, Australia.<br />

Mtuszak G, (2007) Enterprise 2.0 Tales and Trenches”, Information Communications & Entertainment (KPMG<br />

International).<br />

Intermec Technologies Corporation (2007) “Supply Chain RFID: How It Works and Why It Pays”, whitepaper<br />

Intermec Technologies, North America.<br />

O’Reilly T, (2007), What is web 2.0, O’Reilly Media Inc.<br />

Shen, G. and Liu, B. (2010), Research on Application of Internet of Things in Electronic commerce, Third<br />

International Symposium on Electronic Commerce and Security IEEE Society, Beijing Wuzi University, China.<br />

Weinstein R, (2005), RFID: A Technical Overview and Its Application to the Enterprise, IEEE Computer Society,<br />

United States.<br />

Yan B., Huang G, (2008) ,Application of RFID and Internet of Things in monitoring and Anti-counterfeiting for<br />

Products , School of Economics and Commerce, South China University of Technology, Guangzhou, China.<br />

Xin C., (2009), RFID Application in the Retail Industry, Guangdong University of Foreign Studies, 2009 ISECS<br />

International Colloquium on Computing, communication, Control and Management ,China.<br />

307


Modeling the Genetic Schemes of Human Memory<br />

Information in the Process of Production and Reparation of<br />

Knowledge<br />

Fatemeh Mohammadi, Tahereh Mohammadi and Bahram Rezaie<br />

Faculty Member of Islamic Azad University, Shiraz, Iran<br />

faramatn@yahoo.com<br />

tmohammadi@live.com<br />

bahramrezaie@gmail.com<br />

Abstract: In this paper the theory of "genetic scheme of human information" is presented for the first time. This<br />

theory is in based on fifteen years of continuous study and research on related sciences including cognitive and<br />

meta-cognitive psychology, the system of processing information, information pathologies such as amnesia,<br />

information interference, mistakes in remembering information such as the time of tests and exams, neurology and<br />

learning, chemistry and long term memory compounds, electricity physics and electrons movements in data<br />

transferring. To make the theory more comprehensible, a conceptual model on the process of receiving knowledge,<br />

repairing knowledge and finding the relationship between different sciences to each other was created and<br />

presented. This included the use of factor analysis method and path analysis with multiple variable regressions. Its<br />

fitness degree was accepted with 86.12% by a group of 186 Faculty members of Iranian universities. The paper also<br />

offers a conceptual model for understanding the way information system of human mind performs in data saving and<br />

recalling. This model benefits from the results of previous studies and researches on the way memory performs In<br />

terms of neuroanatomy, neuropsychology, physiology, and the system of data processing in the field of<br />

metacognitive Psychology. This model was approved by a group of 97 lecturers in education and neuropsychology,<br />

who have been randomly selected. The related ratio is 85.538. The statistical method employed is factorial and path<br />

analysis. This model depicts the structure of the type of information and the way it is preserved and saved in the<br />

memory. In addition, the model makes it possible to transform the data into other forms which can be recalled and<br />

recognized in a simpler way.<br />

Keywords: information genetic scheme, memory, knowledge generation and reparation, human mind information<br />

system (HMIS), conceptual model, data saving and recalling<br />

1. Introduction<br />

The way human mind performs in saving, processing and recovering information has been studied for<br />

many years by different fields of study including psychology, neurology, and education. The result of<br />

these studies offers models, paradigms, and conceptual frameworks to simplify the understanding of<br />

complexities of human mind performance and his intelligence. (Sternberg, 2006)<br />

These authors’ primary concern was whether a simple conceptual model can show the way information<br />

system of human mind performs, and whether this model can help human to save, process, and recover<br />

information in a better and faster way?<br />

Human being’s dependency on artificial and man-made memories such as portable microcomputers and<br />

electronic note-books in twenty-first century has resulted in mind laziness and malfunction of human<br />

memory. (Mohammadi, 2010)<br />

Therefore, this paper with its emphasis on using the mentioned model, tries to demonstrate how a person<br />

can avoid forgetfulness and mental laziness and enhance his mental capacities. Based on the<br />

introduction presented, the main research questions of this study are as follows:<br />

What are the main components of the conceptual model of Human Mind Information System?<br />

Which model can be proposed to understand HMIS complexities?<br />

What is the degree of validity In university lecturers’ answers to the questionnaires.<br />

2. The significance of the research<br />

Having an appropriate model for the process of remembering and recalling information in information<br />

system of human mind is as important as using models in other experimental disciplines. If the sum of all<br />

endeavors and studies conducted on the performance of human mind, thought, memory and human<br />

intelligence is considered as a whole, then we must be able to create a concrete symbol to represent this<br />

308


Fatemeh Mohammadi et al.<br />

totality. The model presented in this paper tries to show the complexities of HMIS using simple and<br />

understandable symbols.<br />

3. Review of literature<br />

When a child is born he lacks any type of concept or meaning. No sign or symbol of the world makes<br />

sense to him (Berlo 1970, P.178). When we use a word representing a thing or a matter in the presence<br />

of a child, he mysteriously matches the two types of stimuli and then make a relationship between them.<br />

That is because the brain and its neurons are flexible. On the other hand, Berlo shows that meanings and<br />

concepts exist inside humans (Ibid PP 184-5). We know that the organism which maintains these<br />

concepts is memory (Sachs 1967, P437& wallas, 1921, P16).It seems that these concepts lie in the longterm<br />

memory (Hilgard 1975, pp 236-40).<br />

Studies on growth and evolution of memory have been the specific field of Leontiev. Having worked with<br />

normal and retorted children of various ages, he created a device by which the examinee could use<br />

secondary stimuli in order to remind a series of stimuli which were offered by the examiner. Eventually,<br />

Leontiev showed that the process of mediated remembering is long-termed and difficult. This<br />

remembrance is not regular and that there is no evidence showing if the person performs a special<br />

activity to remember previously saved information or not. Luria called this type of behavior "natural<br />

remembering", because she thinks stimulus is remembered through direct affecting process and without<br />

any mediation. But there are times when a person cannot recover and recall the related information by<br />

provided stimulus (Luria, P. 6l).<br />

Following Tief's studies, Morozova believes remembering the stimulus in relation to the response is<br />

difficult for the child. But secondary stimuli can guide the child's chosen response to a right way. She<br />

found out that the rules for mediated remembering are also true for natural remembering. Luria<br />

conducted other research studies in a different setting but with the same substructure. She shows that<br />

memory's activity can twist from a normal form of speech into complex self-folding functions. In 1929,<br />

Luria’s group started an original research on understanding the meaning of stimuli and function of<br />

mediated devices for remembering a series of words and pictures. This group showed that children at<br />

their first stage of growth are not able to create pictorial stimulus which makes the future remembering<br />

possible. But when they grow up, they project effective technique usage. It seems that older children can<br />

organize some of stimuli characteristics in order to facilitate remembering (Ghasem Zade, 2008, P. 85).<br />

In contrast to previous researchers, Vigotesky believes that mental abilities of human being can be<br />

elevated up to a full capacity. He emphasized the qualitative description of organizing information and the<br />

behavior originated from it. As a result, he founded another psychological discipline which emphasizes<br />

the human's socially organized experiences and the structure of his conscious activities. This new branch<br />

has a close relationship with neurology, but it was different from neurology regarding different controlling<br />

areas of psychological functions. Therefore, understanding the abilities of human brain is needed for<br />

neuropsychology or psychology of nervous system (Luria and Cole, 80, P.64).<br />

Vigotesky offered an approach based on his analysis of the structure of psychological function. First he<br />

considered the relationship between preliminary and elevated psychological functions, and then he<br />

determined their mental classification in a normal adult. Vigotesky's approach acted as a model for future<br />

neuropsychology researches (Luria, pp.66-70).<br />

Durkheim believes that basic mental processes cannot be considered as inner-life manifestation or<br />

outcome of natural evolution. His discussion focused on two questions:<br />

Do the content of thought and basic categories used in describing an experience differ in various<br />

cultures?<br />

And, are human’s intellectual functions different in various cultures?<br />

Berlo showed that primitive people followed different rules and functions compared to people in<br />

modern era because primitive individuals did not organize their information enough (ibid, P 73).<br />

In 1980 opinion, intellectual power in primitive cultures basically was not different from individuals living in<br />

nowadays’ technological communities, because the process of thinking, remembering necessary<br />

information, combining them, and creating consequent information has the same logical rules. But other<br />

gestalt psychologists pay attention to the common characteristics of mind in all cultures. They developed<br />

this theory that the principles of understanding and thought are similar worldwide (Werner, H. 1890-<br />

309


Fatemeh Mohammadi et al.<br />

1960). But mental characteristics are not separate from their biological and physiological structures.<br />

Human being’s nervous system has the responsibility of creating harmony between different biological<br />

functions. (Eysenck and Flanangan, 2009).<br />

Human's information is kept in memory. Physiologists found that memory is in Hippocampus area and<br />

they believe that hypo camp has a major role in recording and keeping human information. Hippocampus<br />

is located at the brain cortex. This border is called limbic. Limbic has a role in the setting of feelings,<br />

behavior, motivation and memory due to several links to other parts of nervous system. Hippocampus is<br />

extended in the length of inner edge of temporal boarder. The structure of Hippocampus reflects strata.<br />

Surface of Para Hippocampus has six layers. If we observe Hippocampus from the surface, we face three<br />

layers (out of six layers). These three layers include surface molecular layer which contains nervous<br />

fibers and small scattered neurons, the pyramidal layer which contains several large pyramidal neurons,<br />

and a multishaped layer whose structure is similar to the multishaped layer of other parts of the surface<br />

(Snell, pp.342-5).<br />

This stratified structure led to us to gain a stratified image from the first model of information system of<br />

human brain. Especially where I found that pyramidal layers of Hippocampus are made of round or oval<br />

compact neurons, which create a synapse with dendrites of pyramidal cells of Hippocampus Here, is the<br />

intersection of bringer and sender fibers of information. The study of this section on theoretical<br />

foundations was very important to find out the basic structure of HMIS.<br />

According to this evidences, Hippocampus has a role in converting the short-term memory to long-term<br />

memory. If Hippocampus is damaged, one cannot keep information in his memory for a long time. This<br />

disorder is called Anterograde amnesia. But in this case, the far-off memory which has already been<br />

formed will not be damaged. Damage to the core of Hippocampus makes severe memory disturbances<br />

(ibid P. 347). Despite these facts, Physiologists believe that what is kept in memory is not primary and<br />

raw information but it is often mixed with a meaning whose root has an inner part. (Snell, 2006, p373).<br />

Until the present time, a collection of various hypotheses has been offered in order to describe the<br />

physiological and neurological function of memory. They include, coding of detailed information in a large<br />

collection, -e.g., remembering the details of a trip, physiological coding in the variety of electrochemical<br />

activity of cells and also chemical changes of neuron bonds in message transferring (Mesulam 1981,<br />

p.250). Besides, Asratyan showed that there exists a great correlation coefficient between learning<br />

process and memory. Also, there is a direct relationship between learning and change of protein<br />

composition in the brain (Asratyan, 1983, P.43)<br />

4. Research methodology<br />

The present work is a descriptive research based on modeling the information system. The statistical<br />

population is all faculty members of Educational Science in the universities of Fars Province, Iran), for<br />

estimating the statistical sample we used Z=1.96, Confidence Coefficient = % 95 and d=0.05. Therefore<br />

our sample was n= 86. Since we estimated that about 15% of questionnaires would not be returned, we<br />

increased the number of subjects to 97, in a random basis.<br />

5. Performance of the questionnaire<br />

The study of theoretical literature and effective components on the model showed that these<br />

components, including basic and secondary components, can be primarily classified into five basic<br />

groups:<br />

Philosophy of the model<br />

Objectives of the model<br />

Theoretical foundation and confirmer of the model<br />

The model’s potential for answering and generalizing<br />

The model’s potential to adapt to facts.<br />

For evaluating the suitability of the philosophy of the model three questions, for the objectives of the<br />

model four questions, for the theoretical foundations five questions, for the potential of answering and<br />

generalizing seven questions, and eventually, for the evolution of suitability of the model adaptation to<br />

experienced facts six questions were formed. The discrete method and percentage answers were chosen<br />

since they are more logical and accountable compared to Licret spectrum. Questionnaire’s validity was<br />

confirmed by the experts and through formal validity. We studied twenty six subjects in a pilot test to<br />

310


Fatemeh Mohammadi et al.<br />

obtain their views for future changes in the questionnaire. To evaluate the inner validity of the<br />

questionnaire, test split-method calculation of intensity of relationship between two halves was used<br />

(r=%67.11). To evaluate the validity of relationship power of each question with total score of<br />

questionnaire, dotted two-threaded correlation coefficient was used with the related r value of 0.71.08.<br />

The validity of all questions was checked against Cronbach's Alpha coefficient. The result was α=0.76<br />

which shows the questionnaire was appropriate for final exertion.<br />

6. HMIS model and the findings<br />

The first question of the study is: What are the basic elements of conceptual model of mind information<br />

system?<br />

Theoretical studies show that the elements of HMIS consist of the following parts:<br />

Short-term and long-term memory<br />

Information processing system<br />

Intelligence and the part responsible for discovering the relationship between phenomena<br />

Power of adaptation and compatibility of human brain<br />

Relationship between information type and the speed of memorization and remembrance.<br />

Relationship between the types and amount of received information from the five senses<br />

Meaningfulness of the relationship between new materials and previous information.<br />

Learning methods with emphasis on Bandora, Gestalt, and Azubell.<br />

Nervous system in message transferring (neuropsychology)<br />

Communication system (communication model in message transferring from sender to receiver)<br />

Semantic flexibility in cognitive field<br />

Previous and subsequent information interference system<br />

Learning psychological affections<br />

Ultra cognition and models of self-leading in learning<br />

In answering the second question of the research i.e., which model can be suggested for understanding<br />

the complexities of human mind information system, elicited factors taken from the first question's answer<br />

were entered in the raw and primary model after lecturers' opinions were reviewed, and they were<br />

classified into five basic sections:<br />

Philosophy of the model<br />

Objectives of the model<br />

Theoretical foundations of the model<br />

Potential of the model in answering and generalizing<br />

Potential of the model in adaptation to the experience facts.<br />

7. Philosophy of the model<br />

Philosophy of the HMIS model was evaluated for understanding its complexities by answering three<br />

questions. The result of the analysis of the answers shows that the focus of the model’s philosophy is on<br />

understanding and facilitation of the memorization process and recalling information from long term<br />

memory. This process follows a neuropsychological mechanism which is very complex and its<br />

understanding is very difficult. In order to facilitate understanding of this complex process, the main<br />

structure of the model is based on the type of information introduced to short-term memory and then to<br />

the long-term memory as well as their placement and saving.<br />

To simplify the model, we put the name of the first upper layer in short-term memory, the "current layer".<br />

Current layer is a layer that keeps received and recalled information for a short period of time. This layer<br />

consists of all of information types. Data analysis of answers to the basic component of the model<br />

philosophy showed that %78.62 of statistical sample accepted this philosophy for suggested model of<br />

human mind information system.<br />

311


8. Objectives of the model<br />

Fatemeh Mohammadi et al.<br />

We explain the objectives of the HMIS model by four questions and by regarding components elicited<br />

from the review of the results of research questionnaire and opinion polls of lecturers. We consider the<br />

following objectives for the model:<br />

Understanding the complexities of the performance of memorization and remembering in our<br />

information system.<br />

Engendering meaningful communication for constant saving and quick remembering.<br />

Familiarizing the lecturers with teaching active methods and involving various senses of students for<br />

effective learning.<br />

Arranging the information in the mind, identifying the reasons of forgetting information, and<br />

preventing information forgetfulness.<br />

Statistical descriptive analysis showed that % 89.15 of sample agreed with these model objectives.<br />

9. Theoretical foundations of the model<br />

The current model’s theoretical foundations were evaluated by five questions. The combination of<br />

theoretical studies and acquired results of previous researches as well as data analysis of the<br />

questionnaire and lecturers’ opinions showed that elicited data from the five senses are not saved and<br />

recovered by the same speed and quality. We resulted that, that specific information is remembered<br />

sooner or it is closer to the current layer or simply chemical combinations act more quickly for the<br />

analysis. Therefore, the conclusion is that information is memorized earlier. In other words, they entered<br />

to their special layer of long-term memory sooner. According to the results of sample answers and<br />

background studies, the following order for saving all types of data is offered:<br />

The nearest type of information to the current layer is the information which is received and saved<br />

through experiment or experience. The reason may be the involvement of more senses for learning<br />

or memorizing. For each of the five senses, chemical links opens. Senses of smell touch and taste<br />

function in this layer.<br />

In the second layer, visual and pictorial information exist. As theoretical studies show, the speed of<br />

image memorization and remembering is more than all other types of information received through<br />

hearing and reading. The bulk of received information from visual sense is more than other senses. It<br />

means that %75 of information is received from the sense of sight.(Mohammadi, 2010)<br />

In the third layer, information received through reading and writing is located. The reason for<br />

durability of read and written information compared to heard information and also their careful and<br />

quick remembering can be the use of empirical senses in memorizing and remembering read and<br />

written information .In the case of sighted-talent people while reading, the form and characteristics of<br />

page and written images are facilitating factors which come closer to the tested layer. In a blind<br />

person, the sense of touch makes perception easier and is closer to test and experience layer<br />

compared to heard information.<br />

Finally in the lowest layer the information from the sense of hearing is saved. The bulk of saved<br />

heard information is less than the information received through visual data. Remembering heard<br />

information is also more difficult and memorizing them lasts only for a short time. We know that %75<br />

of heard information can be lost in 24 hours if they are not repeated. (Mohammadi, 2010)<br />

The analysis of answers to questions of the model showed that surface bulk and type of received<br />

information from the five senses are not equal. Therefore, the layers cannot be at the same level<br />

regarding their setting. So, we should design the surface ratio of each layer according to theoretical<br />

foundations. Studies show that on average % 74 of healthy people gain information through sight sense,<br />

%15.67 through sense of hearing, and % 6.9 from reading and writing. %2.999 of the information is also<br />

perceived through experiment and experience; besides, %0.006 of human being’s information is in the<br />

current layer. So, the surface of each layer changes according to the arrangement of layers’ information,<br />

quality and the bulk of each layer. The following simplified model is suggested for better understanding.<br />

The information analysis of the samples' viewpoints about the component of theoretical basis showed<br />

that %86.09 of them accepted the component.<br />

312


Figure 1: Two types of memory<br />

Fatemeh Mohammadi et al.<br />

Current data layer (Short Term Memory)<br />

Long term Memory<br />

Experienced information layer<br />

Seen information layer<br />

Written information layer<br />

Heard information layer<br />

Figure 2: Types of information layers<br />

10. Model’s potential in answering and generalizing<br />

The elements of model’s potential in answering and generalizing were considered by seven questions.<br />

Since theses elements are more important, the number of questions is more than other parts of the<br />

model. Based on the results of information analysis, the samples' viewpoints, and also the answers for<br />

the first research question, we showed with %95 certainty that the bulk of different information types are<br />

not equal for saving & recovering. The more continuous and detailed the parts from large groups, the<br />

easier they can be recalled them. For instance, obtained information from a journey experience has a<br />

great bulk. It consists of so many small fragments and without any attempt; they are saved in the memory<br />

and can be remembered more easily. Yet, a short sentence or a scientist's birth date which is heard via<br />

radio can hardly be recalled. Such information is shown with a big empty circle, because it cannot be<br />

decomposed. Thus, they cannot be recalled easily unless they can appear in another shape or through<br />

association with other information.<br />

In 1986, Mohsenian Rad in an interactive model on man to man relationship presented a meaning unit as<br />

a ball in human mind. According to him and based on the speed of memorizing & remembering data,<br />

‘listened to’ information can be explained by the smallness and roughness of their shapes. The smaller<br />

parts make an apparent shape which move in a limited space, the greater chance of memorizing and<br />

recalling them. In neuropsychology theories, it is said that as all kinds of information are made of<br />

electrochemical energy, they can change from one form to another and potentiality can be re-read and<br />

remembered. So, there should be a way to get to the surface (the current layer). Each group of<br />

information which can enter to the current layer simultaneously and in the form of a group of information<br />

can be remembered more quickly. The combination of the two, displays a complete picture of the model.<br />

Experienced and experimented information layer<br />

Seen information layer<br />

Written and read information layer<br />

Heard information layer<br />

Figure 3: Symbols of any type of information<br />

313


Fatemeh Mohammadi et al.<br />

The analysis of data obtained from the answers of the sample group on "the power of answering and<br />

generalizing of the model" showed that %89.06 of them agree with the mentioned component.<br />

11. Model's power of adaptation to experienced facts<br />

This element of model's power of adaptation to experienced facts was checked by six questions. The<br />

results of the samples’ viewpoints and statistical analysis of their answers to questions indicated that the<br />

given model has %87.06 of adaptation to the experienced facts. %48.91 of faculty members of the<br />

sample group graded the component, appropriate to completely suitable in the information system of<br />

human mind. The model’s adaptation to experienced facts means that by the help of the model, one can<br />

explain the process of learning level discrepancy from one method to another, based on the movement<br />

from one layer to another one. For instance, when somebody is listening to a song or psalm in order to<br />

recite it, he writes it on a piece of paper because by this method, he can recall the order of the psalm<br />

lines by its position on the paper. In other words, instead of just using the Layer responsible for hearing,<br />

he goes to upper layer and uses the Layer which stores read information, written layer and also the layer<br />

that save pictorial information, In fact while he memorizes the song's lines from their position on the<br />

paper, he creates a link between two above layers.<br />

Current data layer (Short Term Memory)<br />

Experienced information layer<br />

Seen information layer<br />

Written information layer<br />

Heard information layer<br />

Figure 4: Information layers format<br />

As people experiment at a particular time, the amount of experienced information is more remembered<br />

and recalled than the heard information. One can conclude that the passage of transferring information<br />

from current layer to other layers and vice versa cannot have the same expanse. For example, the<br />

expansion of experienced information input and output transferring band is much wider than the<br />

expansion of heard or read information transferring band. Similarly, the width of other layers' bands<br />

cannot be the same. Therefore, the suggested model of human's mind information system will reform in a<br />

way that it can adjust itself to the experienced facts. Thus, the channels of transferring information have<br />

different band expansion according to their bulk and type. The theoretical findings based on the existing<br />

literature and the results of faculty members’ viewpoints were approved by %84.91 of faculty members as<br />

Figure 5: Fully Information layers<br />

314<br />

The current information layer<br />

Experienced information layer<br />

Seen information layer<br />

Written and read information layer<br />

Heard information layer


Fatemeh Mohammadi et al.<br />

12. Experimental evidence of model justification<br />

Experiment and research show that permanence of heard information is less than the other types of<br />

information. Thus remembering heard information is more difficult than the other types of information.<br />

This difficulty includes using more energy in recalling, or a need for a stronger motivation to intrigue<br />

information-saving cells to receive and recall the information. Based on the mentioned results we placed<br />

the heard information in the lower layer, far from the current layer. Similarly, here is evidence for the<br />

placement of other types of information in their layers. For example, the information received through<br />

experiments and experiences is rarely forgotten. Pictures are remembered faster than heard information.<br />

The information that are actively dealt with or are repeated successively like a story, remain longer and<br />

more precisely in long term memory.<br />

13. Ratio of the model<br />

The third question in this research is: "What is the validity or ratio of the information system model of the<br />

human mind from the viewpoint of the faculty members?" The result of the information analysis showed<br />

that faculty members’ viewpoints were approved the HMIS model with %85.538 validity.<br />

Table 1: Answer sheet information of sample group according to the ratio degree of the model<br />

Main elements Completely<br />

agree<br />

Agree No<br />

Comment<br />

Opposite Ratio<br />

Model's Philosophy %56 %38 %5 %2 %87.62<br />

Model's Objectives %68 %21 %8 %3 %89.1<br />

Model's Theoretical<br />

Foundations<br />

%63 %29 %7 %1 %86.09<br />

Model's Potential of Answering and<br />

Generalizing<br />

%71 %23 %5 %1 %89.06<br />

Model's potential of Adaptation to Experience<br />

Facts<br />

%61 %28 %9 %2 %84.91<br />

Total View %85.538<br />

Table 2: Descriptive Information of Model Ratio Degree<br />

Total Sample<br />

Model<br />

Man Woman<br />

Mean of Percentage Number Mean of Percentage Number Mean of Percentage Number<br />

Ratio<br />

Ratio<br />

Ratio<br />

85.538 %100 97 81.132 %52.6 51 90.044 %47.4 46<br />

14. Discussion and conclusion<br />

In order to understand the performance of human mind's information system in memorizing and<br />

remembering information, a conceptual model is presented. In this model, the results of previous studies<br />

and researches on the performance of the mind, including neuroanatomy, neuropsychology, physiology,<br />

learning psychology, and processing the system of information in the field of metacognitive were used.<br />

This model is approved by a group of 97 faculty members of education and neuropsychology that were<br />

randomly selected with the ratio of 85.538.<br />

This model shows the structure of information types and the way of locating and saving them in the mind.<br />

It also shows the reason for forgetting heard information among physically and mentally healthy people.<br />

Indent The reason is the form of recognition of heard information in relation to other types of information.<br />

The size and capacity for saving heard information on the one hand and the low speed of remembering<br />

and recalling them from memory on the other hand cause 75 percent of heard information to be forgotten<br />

after 24 hours. This result is compatible with the results of Bach-and-Rita, Zametkin, Raven, and<br />

Raritan's researches. This model also shows how using different senses in learning lead to deeper<br />

saving of information and faster remembering them. Furthermore, this model makes it possible to change<br />

information to other form that can be recognized and remembered simply. This result is also compatible<br />

with Sharma, Orton and Kolb researches result. Also this model can help us to find out the Process of<br />

Production and Reparation of Knowledge.<br />

315


Acknowledgement<br />

Fatemeh Mohammadi et al.<br />

I would like to convey my gratitude to the editor of this article, Maryam Nasseri and my dear family.<br />

References<br />

Bach, Y. Rita (1990) Brain Plasticity as a basis for recovery of function in humans, Journal of Neuropsychologia,<br />

volume 28, pages 547-554.<br />

Berlo, D.K. (1970). The process of communication: an introduction to theory and practice. New York,<br />

Berlo, N. & R. Naugle, (1985)"Case Studies in Cerebral Plasticity", International Journal of Clinical Neuropsychology,<br />

P7<br />

Eysenck, Michael and Cara Flanangan (2009) Psychological Psychology, Translated By: Seifollah Bahary and<br />

Others, (In Persian), Islamic Azad University, Rudhen Branch, Iran.<br />

Finucci, Whitehouse, Isaacs & Childs(1984) Mattis, French & Rapin (1975)Miles,1986, Derivation and Validation of<br />

Reading Disability, Journal of Developmental Medicine and Child Neurology, Volume 26, Pages 143-153<br />

Galaburda AM, Sanides F, Geschwind N.(1978) Human brain. Cytoarchitectonic left-right asymmetries in the<br />

temporal speech region. Arch Neurol, 35(12):812-7.<br />

Galaburda AM, Sherman GF, Rosen GD, Aboitiz F, Geschwind N.(1985) Developmental dyslexia: four consecutive<br />

patients with cortical anomalies. Ann Neurol,18 (2):222-33.<br />

Ghasem Zade, Habibollah (2008) proceeding of Iranian Cognitive Neuropsychology Symposium, (In Persian) Tehran<br />

University. Tehran, Iran.<br />

Hilgard, E. (1975), Psychology background, Translated by Barahani, M. pp 236-40).<br />

Holt, Rinehart and Winston, Inc. Byron, P. Rourke (2005) Neurology of Learning Disabilities: Past and Future,<br />

Journal of learning disabilities volume 28, pages 111-114<br />

Khoda panahi, M.(2007) An Approach to Neuropsychology Research Method(in persion), Allame Tabatabaee<br />

University, Tehran, Iran.<br />

Kolb B. (1989) Brain Development, Plasticity and Behavior. Journal of American Psychologist, Vol. 44, Pages 1203-<br />

1212.<br />

Leontiev Alexis (1959) Problems of Mental Development, Moscow, Akad, Pedag, Nauk.<br />

Luria A. R and Shilawle A. (1980) the Making of Mind, A Personal Account of Soviet Psychology, Edited by Michael<br />

Cole and Sheila Cole, P.64<br />

Mesulam, M. M. (1981) A Cortical Network for Directed Attention and Unilateral Neglect, Annual of Neurology, vol.<br />

10, 309-325<br />

Mohammadi, Fatemeh (2010) human Relationship in organization, (In Persian), Faramatn Publication, Shiraz, Iran.<br />

Mohsenian Rad, Mehdi (1991) Human Communications, (In Persian) Sorush Publication, Tehran, Iran, pp 315-6.<br />

Orton, S. (1928) Specific Reading Disability, Journal of the American Medical Association, Vol 90, Pages 1095-1099<br />

Raven, J. C. (1965) Guide to Using the Colored Progressive Matrix. San Antonio, TX: Psychological Crop.<br />

Reitan, R. M. and Wolfson, D.(1985) the Halstead-Reitan Neurological Test Battery. Tucson, AZ: Neuropsychology<br />

Press.<br />

Richard Snell (2006) Clinical Neuroanatomy, <strong>6th</strong> edition,? Publication, pp.342-5).<br />

Sharma Lata Prem (2006) Learning Readiness, Sarup Teaching Learning Series, No 19. S.T.L.S.<br />

Sternberg Robert, (2006) Cognitive psychology, Thomson, Wadsworth,4th edition, P.43.<br />

Vygotsky, L S. (1978) Mind in Society, Ed, Michael Cole, Vera John Steiner, Sylvia Scribner, and Ellen Souberman,<br />

Cambridge, Harvard University Press.<br />

Werner, H., & Kaplan, E. (1952). The acquisition of word meaning: A developmental study<br />

Zametkin, A. J. et al (1990) Cerebral Glucose Metabolism in Adult with Hyperactivity of Childhood Onset. New<br />

England Journal of Medicine, Volume 323, Pages 1361-1366<br />

316


Copyright Protection for GIS Vector Map Based on Wavelet<br />

Transform<br />

Amer Sedeeq Mustafa<br />

Al-Mustansiriyah University, Baghdad, Iraq<br />

ahea.aur@gmail.com<br />

Abstract: The copyright of GIS vector map has become an increasingly complex topic. The main technologies which<br />

are used for copyright protection are the copyright marking techniques, and these techniques are delegated by<br />

watermarking. In this paper, a new watermarking algorithm for GIS vector map is presented. The magnitude of the<br />

DWT coefficients of vertices extracted from the map is altered to embed the watermark. The selection of the hiding<br />

coefficients will be depended on the threshold value; and a key used to select the exact locations which embed the<br />

watermark. The watermark bits are embedded in coefficients which should satisfy certain conditions. The Insertion of<br />

the watermark in these selected coefficients is done by changing the specific bits of the coefficients like spatial<br />

domain. The watermarks can be blindly detected without the original map and watermark information. The<br />

experimental results demonstrate the high robustness of the proposed algorithm to various attacks like noise<br />

additions etc, and by evaluating the quality of the watermarked map via the WPSNR with respect to the original map,<br />

we show that the watermarked map is very similar to the original map, which indicates that the proposed<br />

watermarking scheme can achieve a better tradeoff between the robustness and the transparency.<br />

Keywords: 2D vector graphics; digital watermarking; wavelet transform; geographic information system (GIS)<br />

1. Introduction<br />

Nowadays, GIS has been widely applied in many fields. The rapid development of computer<br />

communication and the Internet makes it very easy to expediently exchange data via networks. On the<br />

other hand, it also becomes crucial to protect the digital copyright of various digital Medias. GIS data<br />

represents a high material value due to its high cost of acquisition, which is either performed manually or<br />

through analysis of aerial photography Wang (2007), Endoh (2002).<br />

The preparation of maps and services is the first stage involved in the Author module to enable the<br />

display of maps via internet. Each map layer is prepared base on the data and user requirement.<br />

Appropriate symbols and color are assigned to ensure the displaying of the result which fulfils the criteria<br />

and features required. Next, the map layers will be compiled and stored in a service to ease users<br />

accessing to the application server. With the emergence of digital maps, copyright protection of maps<br />

encountered a new challenge. New technology should be developed to solve this problem. Digital<br />

watermarking technology provides a new solution for digital map copyright protection .Despite the fact<br />

that the protection of vector images can be of great importance in many applications; copyright protection<br />

of vector images still drew little attention within the watermarking fields. There are only a few published<br />

works done on watermarking vector images. Indeed, the research on vector map digital watermark has<br />

just started and authenticity of a vector image is becoming a hot research area Busch (2004).<br />

M. Sakamoto et al proposed a watermarking scheme by modifying the location relationship of vertices<br />

Takashima (2000). Ohbuchi et al presented several algorithms for 2D vector map watermarking. In<br />

spatial domain, a scheme based on correlation detection was proposed in Endoh (2002). Micheal Voigt<br />

and Christoph Busch proposed an algorithm which can embed multibit data based on direct sequence<br />

spread spectrum and modified correlation detection Busch (2003). These algorithms of watermarking are<br />

belong to spatial domain. The watermarking algorithm for vector maps in DFT domain was proposed by<br />

N. Nikolaidis, I. Pitas, and V. Solachidis Solachidis (2000), Pitas (2000) they proposed a blind<br />

watermarking scheme which embeds a single bit into a polyline by modifying the discrete Fourier<br />

coefficients of polyline’s coordinate sequence.<br />

2. Digital watermarking<br />

Digital watermarking is the process of embedding information into digital multimedia content such that the<br />

information (which we call the watermark) can later be extracted or detected for a variety of purposes<br />

including copy prevention and control. Digital watermarking has become an active and important area of<br />

research, and development and commercialization of watermarking techniques is being deemed<br />

essential to help address some of the challenges faced by the rapid proliferation of digital content Wang<br />

(2009).<br />

317


Amer Sedeeq Mustafa<br />

The detection and analysis of digital information can ensure the integrity and reliability of the digital<br />

information; therefore serve as a means to protect the rights and interests of the author. Robust<br />

watermark refers to the watermark’s ability to resist regular signal process and malicious attacks, it is<br />

mainly used to confirm the author, creator, copyright owner, distributor, and authorized consumer of the<br />

digital media. Fragile watermark is a class of watermark which is less robust. It is mainly used to protect<br />

the authenticity and integrity of the document Solachidis (2000).<br />

3. Wavelet transform<br />

Wavelet analysis has been turned into research hot spot in the signal processing in recent years because<br />

it is a powerful analysis tool. The wavelet theory has good performance and it is been widely applied in<br />

voice signal processing, image analysis, data compression and etc.. Eui (2002). With the wide application<br />

of wavelet technology, watermark technology based on wavelet has gained more and more attention.<br />

Wavelet transform can get image multi-scale details step by step via flexing and parallel moving<br />

operation. As scale becomes smaller, every part gets more accurate, and ultimately all image details can<br />

be focalized accurately. If planar discrete wavelet analysis is made to input image, it will produce three<br />

high frequency parts (HL, LH and HH) and one low frequency part (LL). The low frequency part can<br />

denote optimal approach to original image. The majority of image energy concentrates in here.<br />

4. Vector map<br />

The vector map represents real-world features as strings of x,y pairs representing spatial information<br />

about the features. A point is a single pair of values with an x value representing the longitude of the point<br />

in some coordinate space and a y value representing the latitude of the point. Linear features may consist<br />

of a set of connected points or some mathematical function describing the beginning point of the feature<br />

and the formula that constructs it. Polygon features are closed loop of X-Y coordinate pairs, last pair are<br />

the same. The example of vector map is shown in Figuer 1.<br />

Figure 1: Example of vector map<br />

5. The proposed algorithm<br />

The block diagram of the proposed watermarking algorithm is shown in Figuer 2.<br />

5.1 Vertex extraction and extension<br />

All vertices are extracted from the 2D vector graphics V, where N means the number of vertices. Let {Vk }<br />

be vertex sequence, Vk = (xk , yk ) means vertex coordinates, xk and yk means longitude and latitude<br />

coordinates respectively. The number of vertices may be not in multiples of 32, so extension of vertices in<br />

a minimum scope is proposed. The increased number of vertices can be calculated by:<br />

M = min {x | (N + x) mod 32 = 0} (x =1, 2….) (1)<br />

318


Amer Sedeeq Mustafa<br />

Figure 2: Block diagram of the proposed watermarking algorithm<br />

The value of increased vertex coordinates is the same as the final vertex coordinate. After this step, the<br />

extracted vertices will be sorted in vector of one dimension as follow:<br />

V1 V2 ……………………………….. V32<br />

Each vertex Vk is composed of the x and y coordinates, so each group has 64 coordinates, and then<br />

structure 8 × 8 matrix based on these 64 coordinates can be constructed as follow:<br />

5.2 Wavelet transform<br />

x1 y1 x2 y2 x3 y3 x4 y4<br />

..<br />

..<br />

..<br />

..<br />

..<br />

..<br />

.. .. .. .. .. ..<br />

.. .. .. .. .. .. x32 y32<br />

The proposed method for hiding watermark is based on the Discrete Wavelet Transform (DWT). The<br />

wavelet transform is used to generate watermark and decompose the 8 × 8 matrix into regions to help in<br />

selecting the suitable locations to hide the watermark in vector map.<br />

We applied one level discrete wavelet decomposition on the 8 × 8 matrix. The Haar wavelet basis is<br />

chosen due to its simplicity. The 2×2 Haar matrix that is associated with the Haar wavelet is:<br />

319


Amer Sedeeq Mustafa<br />

⎡1<br />

1 ⎤<br />

H2= ⎢ ⎥<br />

2)<br />

⎣1<br />

−1⎦<br />

The DWT separates 8 × 8 matrix into a lower resolution (LL) as well as horizontal (HL), vertical (LH) and<br />

diagonal (HH) detail components. The 2×2 Haar matrix is applied in the following way:<br />

G'(i, j)= H2 [G(i, j)] (3)<br />

Where G(i, j) refers to matrix values and will be modified to G′ (i, j) wavelet coefficients matrix after DWT.<br />

5.3 Watermark generation<br />

In this step we will generate watermark bits from wavelet coefficients matrix G ′ (i, j), and generate an 8 ×<br />

8 binary matrix A according to wavelet coefficient matrix G′ . The values of matrix A can be defined as :<br />

⎧ 1 G'(i,<br />

j) > M<br />

⎨<br />

A(i, j)= ⎩ 0 G'(i,<br />

j) ≤ M<br />

(4)<br />

Where M is the mean value of G′ and i,j=1,2…8. Watermark bits are generated as follows:<br />

P= A(1, 1) ⊕ A(2, 1) ⊕ ………….. A(8, 1) (5)<br />

Where ⊕ refers to XOR.<br />

5.4 Watermark embedding<br />

After separates 8 × 8 matrix into four channels LL, HL, LH, and HH. Since LL coefficient modification can<br />

easily lead to major changes in vertex coordinates thereby reducing the transparency of the watermark,<br />

while the HH coefficients are too sensitive, we choose the HL and LH regions of the coefficients. The<br />

hiding coefficients will be selected depending on the threshold value and a secret key. This key is used to<br />

select the exact locations which embed the watermark.<br />

The threshold value is calculated for HL and LH regions by using the following equation Eui (2002):<br />

T=2 (log2max (Ci))-1 (6)<br />

Where Ci represents the largest integer coefficient HL and LH regions. The watermark is embedded only<br />

for the selected coefficients.<br />

The watermark bits are embedded in coefficients that are in HL and LH subbands. The host coefficients<br />

should satisfy the following condition:<br />

|Coefficient Magnitude | > Threshold value.<br />

Inserting the watermark in these selected coefficients is done by changing the specific bit of the<br />

coefficients like spatial domain. We use inverse DWT transform to produce watermarked matrix, and then<br />

transform the matrix into one-dimensional sequence and remove extended vertices to get the<br />

watermarked vertex coordinates. Then get the watermarked map.<br />

5.5 Watermark extraction<br />

Watermark extraction is the inverse process of watermark embedding, only need to know the threshold<br />

value and the secret key. The steps are as follows:<br />

Extract all the vertices of test vector map and then extend to set every group 32 vertices.<br />

8×8 matrix is generated.<br />

Make DWT transform on the matrix.<br />

320


Amer Sedeeq Mustafa<br />

Generate an 8 × 8 binary matrix A′ according to this DWT coefficient matrix as in Eq.(4). Extract<br />

watermark bits p′ as follows:<br />

p′ = A′ (1, 1) ⊕ A′ (2, 1) ⊕ ……….. A′ (8, 1).<br />

Calculate threshold value for HL and LH regions as in Eq. (6).<br />

According to the secret key and the threshold value the selected hiding locations where watermark<br />

embedded can be determined and watermark bits W’ can be gotten.<br />

Since we select two middle DWT coefficients (HL and LH) to embed watermark repeatedly, we obtain<br />

W and W . If p′ = W and p′ = W , then this map is copyrighted.<br />

'<br />

1<br />

'<br />

2<br />

'<br />

1<br />

6. Experimental results<br />

'<br />

2<br />

In order to evaluate the proposed watermarking algorithm, two vector maps are used in Figuer 3.<br />

Figure 3: Watermarked vector maps (map1 and map2)<br />

So as to test the robustness of the proposed algorithm, we try to apply different kinds of operation to the<br />

watermarked data, and extract the watermarking information after that. After extracting watermark, we<br />

use similar measurement of the extracted and the referenced watermarks as validation; it can be defined<br />

by the Normalized Correlation (NC). The formula which calculates similarity is defined as:<br />

∑<br />

∑<br />

*<br />

* 2<br />

NC= ( W . W ) / ( W ) , k=1,2,…..L (7)<br />

k<br />

k<br />

k<br />

k<br />

k<br />

*<br />

Where, W and W are referenced and extracted watermarks. L is length of watermark bits.<br />

k<br />

k<br />

Meanwhile, the visual quality of watermarked and attacked images is measured by using the Peak Signal<br />

to Noise Ratio (PSNR). A ’PSNR’ derived from digital image fidelity evaluating is presented in this paper,<br />

which is used as a measure of the map quality:<br />

PSNR( dB)= 10log10<br />

2 ⎡max<br />

⎤<br />

∀(n,<br />

m) f (i, j)<br />

⎢<br />

⎥<br />

⎢⎣<br />

MSE ⎥⎦<br />

Where MSE is the mean square error:<br />

MSE=<br />

∑∀(n, m)<br />

(fw(i,<br />

j) − f(i, j))<br />

nm<br />

2<br />

321<br />

(8)<br />

(9)


Amer Sedeeq Mustafa<br />

where n, m are the dimensions of the input map and f, fw are the original and the watermarked maps.<br />

However, PSNR declines from the perceived subjective quality because the Human Visual System (HVS)<br />

does not correlate well with the square of the error. For this reason, the Weighted Peak Signal to Noise<br />

Ratio (WPSNR) that takes into account the properties of human visual system is also used as follows<br />

Zhang (2009):<br />

WPSNR(dB)=10 log10<br />

2 ⎡max<br />

⎤<br />

∀(n,<br />

m) f (i, j)<br />

⎢<br />

⎥<br />

⎢⎣<br />

MSE * NVF ⎥⎦<br />

2<br />

2<br />

φσ ( i,<br />

j))<br />

σ ( i,<br />

j)<br />

Where Noise Visibility Function NVF (i, j) =1/(1+ x , x denotes the local variance of the<br />

map in a window centered on the vertex with coordinates (i, j).<br />

2<br />

2<br />

θ is a tuning parameter corresponding to the particular map and is given as θ =D/ σ , where σ is<br />

the maximum local variance for a given map and D is an experimental value, range from 50 to 100. Table<br />

1 shows the watermark extracting results of the proposed algorithm.<br />

Table 1: Result of test<br />

Attack<br />

x max<br />

Result table for illustration the performance of watermarking algorithm<br />

Map 1 Map2<br />

PSNR<br />

WPSNR<br />

20% JPEG compression 34.1 36.7 0.9720 27.5 35.2 0.9654<br />

40% JPEG compression 32.2 33.8 0.9684 27.5 35.2 0.9521<br />

Salt and Pepper Noise 27.75 30.3 0.94 27.5 35.1 1.00<br />

Gaussian lowpass filter 44.96 47.9 0.988 41.2 43.8 0.986<br />

1/4 cropping 11.11 13.5 0.862 10.4 18.0 0.861<br />

7. Conclusions<br />

In this paper, we propose a watermarking algorithm which can be used for GIS vector map. Experimental<br />

results show that the degradation by embedding the watermark is too small to be visualized. In addition,<br />

the robustness of this algorithm is strong enough to some attack like noise, compression, and cropping.<br />

Therefore, the proposed watermarking scheme can achieve a better tradeoff between the robustness and<br />

the transparency.<br />

Future work will focus on making the watermarks robust to more attacks and optimizing the algorithms to<br />

provide higher capacity and efficiency.<br />

References<br />

Nikolaidis, N, Pitas, I and Solachidis, V. (2000) “Fourier Descriptors Watermarking of Vector Graphics Images”, Proc.<br />

of the International <strong>Conference</strong> on Image Processing, vol.3, pp.10-13.<br />

Ohbuchi R, Ueda H, and S. Endoh (2002) “Robust Watermarking of Vector Digital Maps”, Proc. of IEEE International<br />

<strong>Conference</strong> on Multimedia and Expo (ICME02), Lausanne, Switzerland, pp.577-580.<br />

Qingtang SU, Xianxi LIU, Steve Zhang (2009) ”Quality Evaluation of Digital Image Watermarking”, Fifth International<br />

<strong>Conference</strong> on Intelligent Information Hiding and Multimedia Signal Processing.<br />

Sakamoato, M, Matsuura, Y and Takashima, Y (2000) “A Scheme Of Digital Watermarking for Geographical Map<br />

Data”, Proc. of the Symposium on Cryptography and Information Security, Okinama, Japan, pp.26-28.<br />

322<br />

NC<br />

PSNR<br />

WPSNR<br />

(10)<br />

x max<br />

NC


Amer Sedeeq Mustafa<br />

Sung. K. Je, Cheol. K. Kim, and Eui (2002). Y. Cha “An Adaptive Watermarking Algorithm Using Wavelet Transform”,<br />

Neural Network and Real-World Applications Lab, Department of Computer Science, Pusan National University,<br />

Gumjung-gu, Pusan 609-735, Korea.<br />

Solachidis, V, Nikolaidis, N and Pitas, I. (2000) “Watermarking Polygonal Lines using Fourier Descriptors”, Proc. of<br />

the IEEE International <strong>Conference</strong> on Acoustics, Speech and Signal Processing, Istanbul, Turkey, vol. IV,<br />

pp.1955-1958.<br />

Voigt, M and Busch, C. (2003) “Feature-based Watermarking of 2D-vector Data”, Proc. of the SPIE, Security and<br />

Watermarking of Multimedia Content, Santa Clara, USA, vol.5020, pp.359-366.<br />

Voigt, M, Yang, B and Busch, C. (2004) “Reversible Watermarking of 2D-vector Data”, Proc. of the Multimedia and<br />

Security Workshop on Multimedia and Security, Magdeburg, Germany, pp.160-165.<br />

Xia M. Niu, ChengY. Shao, and Xiao T. Wang (2007) “GIS Watermarking: Hiding Data in 2D Vector Maps”, Studies in<br />

Computational Intelligence, vol.58 , pp.123-155.<br />

Zheng, Liangbin, Wang, Yulu Jia, Qun. (2009) “Research on Vector Map Digital Watermarking Technology”, Proc. of<br />

the IEEE Education Technology and Computer Science. ETCS '09. First International Workshop.<br />

323


Evaluation of IT Investment Methods and Proposing a<br />

Decision Making Model<br />

Shirin Nasher, Mehrdad Kalantarian, Ahmad Akbari, Ali Suzangar, Mohammad<br />

Kajbaf and Negar Madani<br />

Infoamn IT Consultancy CO., Tehran, Iran<br />

shirin_nasher@vu.iust.ac.ir<br />

mehrdad_kalantarian@vu.iust.ac.ir<br />

akbari@iust.ac.ir<br />

a.suzangar@infoamn.com<br />

m.kajbaf@infoamn.com<br />

n.madani@infoamn.com<br />

Abstract: Information technology investment decision making is one of the significant issues. Since the IT<br />

investment evaluation is not just based on direct and tangible factors and many other intangible and indirect<br />

qualitative criteria influence this evaluation. Generally, there are two different approaches in evaluation methods with<br />

their own advantages and disadvantages: tangible methods such as Discounted Cash Flow, Net Present Value,<br />

Information Economics, etc. and intangible methods such as Value Analysis, Multi Objective Multi Criteria, Critical<br />

Success Factors, etc. But a more effective and precise road map is to guide decision makers to choose an<br />

appropriate multi criteria model that consider both tangible and intangible factors together. In this paper by literature<br />

review of these mentioned methods, various tangible and intangible factors were determined from different academic<br />

papers and practitioner resources and then were classified in two domains and a number of sub-domains. In order to<br />

obtain complete and applicable criteria, five reduction factors were defined, i.e. clarity, completeness, non–<br />

redundancy and operationality. Then this criteria list was delivered among a number of mangers and information<br />

technology specialists. According to the given answers, a new criteria list was obtained with eliminating nonapplicable<br />

criteria. As a consequence, to assess the importance of each criterion for creating the model, the rating<br />

scores, one to five were defined and added to the new lists. Then these new criteria lists were conducted among<br />

CIO, CEO, CFO and other related specialists in different Iranian companies to customize these criteria according to<br />

their business strategies and requirements. Score one represents not important and five represents very important.<br />

The results are assumed as the minimum level of criteria with maximum coverage in different information technology<br />

projects. In the next step, based on the results, we developed an analytic hierarchy decision making model. The<br />

results of this research indicate that this model is applicable and can be easily expanded by aggregating new sub<br />

criteria to be customized for different IT investment evaluations.<br />

Keywords: IT investment, evaluation of IT project, intangible benefits, decision making model, economic factors,<br />

rational decision making<br />

1. Introduction<br />

Information technology is actively applied in various segments of society that has a strong impact on the<br />

global performance of the businesses. In the past few years, the rapid growth of IT investment is one of<br />

the most incredible issues in different business units and organizations. On the other hand, its adoption is<br />

based on tangible and intangible aspects. But there is a strong persistency to concentrate on tangible<br />

aspects, but the fact is that intangible factors often are the most important one associated to investment.<br />

Indeed, it is not easy for administrations to evaluate the real return of IT investment, so convincing<br />

organization chief managers to allocate their budget to information technology projects is difficult. Even<br />

though organizations are eager to spend considerable amount of time and money to select appropriate IT<br />

project, but they may not include all relevant criteria in evaluating the IT investment (Wu and Ong<br />

2008).Hence even nowadays aligning IT investment strategies with business policies is an important<br />

issue in organizations (Borenstein and Baptista Betencourt 2005), (Goh and Kauffman 2005), (Joseph<br />

Wen, Yen and Lin 1998).<br />

Although there are different methods and tools developed for IT investment evaluation, but the majority<br />

only take easily measured financial aspects into account without any attention to strategic or operational<br />

aspects. Among these methods, the multi criteria multi objective method combines various criteria in<br />

general specific and applied to a unique situation. Hence the existence of a model based on the validated<br />

tangible and intangible criteria, independent from the context related to the specific IT investment seems<br />

very important to support a decision (Chen, Zhang and Lai 2009), (Schniederjans and Hamaker 2003),<br />

(Schniederjans , Hamaker and Schniederjans 2004).<br />

324


Shirin Nasher et al.<br />

The main goal of this study is to identify and structure a set of relevant and validated minimal measurable<br />

set of criteria and sub criteria to formulate a hierarchy decision making model for a specific IT investment.<br />

These criteria were raised from different research and practitioner resources and validated through the<br />

evaluation of specialist to support and develop decision makers confidence in the selection of the most<br />

suitable project for a particular case (Goldstein, Katz and Olson 2003), (Mahmood and Szewczak 1999).<br />

In this paper, first in the literature review section, different tangible or intangible IT investment evaluation<br />

methods will be introduced. Then the advantages and disadvantages of these methods will be compared<br />

to each other.<br />

Second, the methodology of this research will be described in the consequence. At last in the conclusion<br />

the benefits of the proposed model will be presented.<br />

2. Literature review<br />

The best IT investments are those which help to maximize the value of the firm. They also contend that to<br />

maximize the value of the firm, IT investment decisions need to be able to maximize IT benefits while<br />

minimize IT risks (Joseph Wen, Yen and Lin 1998).<br />

There are three reasons why being familiar with IT benefits and risks is important. First, some IT benefits<br />

are lost through inappropriate management, while some are lost because they are not recognized by the<br />

management in the early IT planning stage. Second, it is important to identify the benefits to be measured<br />

prior to selecting applicable evaluation methods. This is because some methods are suitable for<br />

evaluating tangible benefits while others are more suitable for intangible benefits. Finally, the recognition<br />

of risk as an important component in IT investment decision making has long been recognized (Epstein<br />

and Rejc 2004-2005), (Joseph Wen, Yen and Lin 1998).<br />

2.1 Information technology benefits and risks<br />

In general, the benefits of IT investments can be classified into five broad classes, the purpose of which<br />

is to (1) increase productivity and operating process performance; (2) facilitate management support; (3)<br />

gain competitive advantages; (4) provide a good framework for business restructure or transformation<br />

and (5) provide clearance of expenditures (Epstein and Rejc 2004-2005).<br />

IT has provided many benefits to corporations over the years. Since we are living in a global information<br />

society with a global economy, which is increasingly dependent on the creation, management and<br />

distribution of information resource. However, investments in information technology are subject to higher<br />

risks than any other capital investments for several reasons. First, their components are comparatively<br />

fragile. Second, information systems are likely to be the target of disgruntled workers, protester, and even<br />

criminals. They can also fall in the hands of the competitors. Finally, the decentralization of information<br />

systems and the use of distributed processing have increased the difficulty of design, development,<br />

management, and protecting information systems. IT risks are classified into two general classes: (1)<br />

physical risks; and (2) managerial risks (Epstein and Rejc 2004-2005).<br />

2.2 Information technology investment evaluation methods<br />

Given countless existing methods, a broader framework for categorizing and understanding evaluation<br />

techniques seems highly desirable. Such a framework by dividing IT evaluation approaches into two<br />

broad categories based upon their underlying assumptions: objective/rational or subjective/political. In the<br />

objective/rational category, they further divided the objective/rational category into two zones: efficiency<br />

i.e., doing things correctly and effectiveness i.e., doing the correct things. The subjective/political<br />

category described as the understanding zone i.e., discovering why things are done (Tuten 2009).<br />

Traditional IT evaluation practice operates from an objective/rational point of view, focusing on the<br />

efficiency and effectiveness of solutions. Such evaluation approaches are grounded in a positivist<br />

epistemology—an epistemology that, when applied to this context, holds that information systems are<br />

inherently objective and rational. Therefore, practitioners should evaluate information systems using<br />

objective/rational methods (Tuten 2009).<br />

Overall, researchers have tended to describe traditional evaluation methods as formal, overt, ritualistic,<br />

mechanistic, quantitative, and/or prescriptive in their efforts to determine the costs, benefits, and risks<br />

associated with IT investments (Tuten 2009).<br />

325


Shirin Nasher et al.<br />

Nevertheless, researchers have suggested that formal evaluation frequently fails to be undertaken with<br />

rigor and is completely avoided by practitioners in many cases. In a recent study IT evaluation practices<br />

in <strong>European</strong> companies, researchers found that only one third of the organizations surveyed conducted<br />

formal evaluations. Yet, when organizations perform IT evaluation, they tend to employ traditional<br />

methods that hold considerable legitimacy with executives and managers. This finding found that<br />

quantitative evaluation methods were widely used by the organizations conducting formal evaluations<br />

(Tuten 2009). Table1 shows these methods based on mentioned categorize.<br />

As Table1 shows, these rational/objective effectiveness methods may be subcategorized into one of<br />

three groups of methods: economic, non-economic, and hybrid.<br />

Table 1: The information technology investment evaluation methods (Tuten 2009)<br />

Economic<br />

Discounted Cash Flow<br />

Cost Benefit Analysis<br />

Payback Period<br />

SEASAME<br />

Return on Management<br />

Return on Investment<br />

Options Theory<br />

Risk/Sensitivity Analysis<br />

Effectiveness Zone<br />

Non-economic<br />

User Information<br />

Satisfaction<br />

Hybrid<br />

Balanced Score Card<br />

Critical Success Factors<br />

Information Economics<br />

Multi-Criteria Approaches<br />

Value Analysis<br />

Efficiency Zone<br />

Simulation<br />

TQM/Software<br />

Metrics<br />

2.2.1 Economic methods<br />

In particular, the researcher discussed each of the following widely cited methods: Discounted Cash Flow<br />

(DCF) techniques, Cost/Benefit Analysis (CBA), payback period, Systems Effectiveness Study and<br />

Management Endorsement (SESAME), Return on Management (ROM), Return on Investment (ROI),<br />

options theory, and risk sensitivity analysis (Tuten 2009).<br />

2.2.2 Non-economic method<br />

In the rational/objective literature stream, techniques for measuring user satisfaction, particularly the User<br />

Information Satisfaction (UIS) method, provide the notable exception (Tuten 2009).<br />

2.2.3 Hybrid methods<br />

Hybrid approaches may utilize financial/economic factors and/or non-economic dimensions to evaluate<br />

information systems. All of the following methods have been associated with the rational/objective stream<br />

of IT evaluation techniques. These approaches vary considerably with respect to their degree of apparent<br />

objectivity, as demonstrated by either their reliance on quantitative measures or empirically observable<br />

outcomes. For example, in practice Information Economics relies heavily on their quantitative enhanced<br />

ROI metric. In contrast, Critical Success Factors (CSF) method utilizes a dialogic approach to uncover<br />

executives’ explicit and implicit goals and objectives. In this sense, the term hybrid provides an apt<br />

description for this group’s diversity of methods and measures. These methods are such as Balanced<br />

Score Card (BSC), Critical Success Factors (CSF), Information Economics (Parker and Besnson 1988)<br />

and etc. (Tuten 2009).<br />

Note that, the focus of this paper is on the effectiveness zone methods, so it has preferred to eliminate<br />

describing other zone. Since there are differences between these more applicable methods, Table2 has<br />

summarized the previous discussions based on methods attributes.<br />

In Table2, the IT benefit factors column shows that which of the mentioned above methods concentration<br />

is on tangible or intangible criteria. Therefore these benefit factors can determine the criteria that are<br />

required for IT investment evaluation and decision making lastly.<br />

326


Shirin Nasher et al.<br />

Table 2: Differences among information technology investment evaluation methods (Joseph Wen, Yen<br />

and Lin 1998)<br />

Evaluation<br />

category<br />

ROI<br />

Model/procedure<br />

examples<br />

NPV, DCF, pay-back<br />

period formulas<br />

Measures of IT<br />

benefits factors<br />

tangible<br />

Measures of<br />

IT risks<br />

discount<br />

rates,<br />

surrogate<br />

measures<br />

CBA cost/benefits formulas tangible factors same as ROI<br />

ROM<br />

IE<br />

MOMC<br />

VA<br />

CSF<br />

Productivity based<br />

formulas<br />

same as ROI<br />

supplemented with<br />

ranking and scoring<br />

math models and<br />

multistage iterative<br />

processes<br />

multistage, evolutionary<br />

process<br />

multistage,<br />

evolutionary process<br />

RO Multistage process<br />

PA Financial models<br />

Delphi<br />

approach<br />

3. Methodology<br />

multistage,<br />

evolutionary process<br />

tangible, labor<br />

value-added as<br />

intangible<br />

tangible and<br />

some intangible<br />

tangible and<br />

intangible<br />

tangible factors<br />

user’s surrogate<br />

measures<br />

tangible and<br />

intangible<br />

factors<br />

measures for<br />

cost<br />

savings<br />

user’s surrogate<br />

measures<br />

Not<br />

addressed<br />

surrogate<br />

measures,<br />

risks with<br />

ranking and<br />

scoring<br />

several<br />

measures<br />

of utility and<br />

risks<br />

not<br />

addressed<br />

user’s<br />

surrogate<br />

measures<br />

surrogate<br />

measures<br />

for risks<br />

and costs<br />

direct<br />

measures<br />

of risks<br />

user’s<br />

surrogate<br />

measures<br />

Major<br />

advantages<br />

mainly<br />

quantitative<br />

focus on<br />

efficiency<br />

mainly<br />

quantitative<br />

focus on<br />

effectiveness<br />

mostly qualitative<br />

measures of<br />

efficiency<br />

qualitative and<br />

quantitative<br />

measures<br />

mainly<br />

quantitative,<br />

multiple and<br />

conflicting<br />

objectives<br />

tangible factors<br />

intangible<br />

factors,<br />

centered on<br />

effectiveness<br />

many intangible,<br />

centered<br />

on effectiveness<br />

higher efficiency<br />

tangible and<br />

intangible<br />

factors<br />

Major limitations<br />

no intangible,<br />

reliance<br />

on accounting<br />

data<br />

surrogate<br />

measures<br />

for intangible<br />

factors<br />

limited<br />

quantitative<br />

measures,<br />

assumptions<br />

hard to meet<br />

major<br />

simplifying<br />

assumptions<br />

and<br />

models<br />

relatively new<br />

in MIS, still in<br />

development<br />

prototyping,<br />

need<br />

several<br />

revisions to<br />

final results<br />

highly<br />

qualitative<br />

process<br />

highly<br />

subjective and<br />

qualitative<br />

mainly<br />

quantitative<br />

highly<br />

qualitative<br />

In this paper, IT investment decision making problem has been presented. The methodology of problem<br />

solving with the purpose of building the model to evaluate IT investments would be described here. This<br />

methodology has three steps that would be explained as follows:<br />

Gathering and reducing criteria<br />

Weighting<br />

Proposing model<br />

3.1 Gathering and reducing criteria<br />

This stage has comprised of two steps. At first by reviewing pervious research studies concerning IT<br />

investment evaluation, many of the criteria and sub criteria were identified (Renkema and Berghoutb<br />

1997), (Kraemer and Dedrick 1994), (Joseph Wen, Yen and Lin 1998), (Henderson and Venkatraman<br />

1990). More than 250 criteria were enlisted in an excel file that were concerned both tangible and<br />

intangible factors. After that these criteria were classified into four domains in the separated excel sheets,<br />

327


Shirin Nasher et al.<br />

i.e., operational, tactical, strategic and risk. Each domain has its own criteria and relative sub criteria. The<br />

definitions and characteristics of the criteria were described in each sheet. The first sheet was considered<br />

for the brief description about this research.<br />

Second, because this raw list has many unrealistic and non-applicable criteria, a reduction process was<br />

conducted to identify clear and operational criteria list. Therefore, reduction factors were defined as<br />

follows: the clarity, completeness, non–redundancy and operationality. In the consequence, these factors<br />

have been added for each criterion in the new columns with check boxes to that excel sheet.<br />

This file has delivered among ten managers and information technology specialists. Based on the<br />

gathered results, these actions were done:<br />

1. If the most specialists had voted that the criterion is not clear, it must be defined clear.<br />

2. If the most specialists had voted that the criterion is not complete, it must be completed.<br />

3. If the most specialists had voted that the criterion is redundant, it must be eliminated.<br />

4. If the most specialists had voted that the criterion is not operational, it must be changed or at last<br />

eliminated.<br />

After applying reduction factors to the set of 250 criteria and these actions, several criteria were<br />

eliminated and several were changed or completed. At last the new minimal measurable list of criteria<br />

was obtained. This new list has two domains, i.e. “strategic” and “operational and tactical” with eleven<br />

criteria and forty two sub criteria. A brief description of each eleven criteria is as follows:<br />

1. Strategic alignment: This criterion is related to this crucial point that which investment can align<br />

with organizational strategy and also the possibility of IT providing opportunities for businesses.<br />

2. Competitive advantages: It refers to the organization competitive position among market and<br />

supplier and other related items that mentioned in this criterion.<br />

3. Organizational productivity: It is about the organizational productivity improvement.<br />

4. Market assessment: This criterion deals with the pervious background of the specific IT project<br />

and mentions its former customers and their credit in the market. Also, investigates the ability of<br />

project in the market anticipating.<br />

5. Compliance: It seeks to identify the compliance between proposed investment with the current<br />

internal or external requirements and standards.<br />

6. Quality: It tends to focus on the IT performance characteristics and the impact of it on the services<br />

improvement to support business processes.<br />

7. Security: This criterion and its related items refer to three main aspects (Confidentiality, Integrity<br />

and availability) of security in information systems.<br />

8. Flexibility and compatibility: This criterion and its related items refer to assess program<br />

reliability.<br />

9. Cost: It is related to the different costs associated with a particular investment.<br />

10. Satisfaction: It considers external customers or internal user’s satisfaction about the new<br />

proposed investment.<br />

11. Effects: This criterion tends to focus on the effects of investment on the internal organizational<br />

culture and communications.<br />

The primary model based on these minimal measurable validated criteria was shown in figure1.<br />

3.2 Weighting<br />

In this step to determine domains, criteria and sub criteria importance in order to proposing a model, 5<br />

importance degrees were defined in new lists for each criterion i.e., very important, important, relatively<br />

important, less important and not important. Then these new lists were conducted among twenty of IT<br />

experts with minimum 10 years professional experience in Iran. They determined importance degree for<br />

each criterion based on their own organizational strategies and business requirements. The scores one<br />

to five were defined that score one represents not important and five represents very important. After<br />

answer sheet gathering, first the stability of given answers was calculated by SPSS. It was 91.2 percent<br />

which indicates that IT expert’s answers were at minimum deviation.<br />

328


Shirin Nasher et al.<br />

Figure1: The multi criteria model for IT investment evaluation<br />

3.3 Weighting<br />

In this step to determine domains, criteria and sub criteria importance in order to proposing a model, 5<br />

importance degrees were defined in new lists for each criterion i.e., very important, important, relatively<br />

important, less important and not important. Then these new lists were conducted among twenty of IT<br />

experts with minimum 10 years professional experience in Iran. They determined importance degree for<br />

each criterion based on their own organizational strategies and business requirements. The scores one<br />

to five were defined that score one represents not important and five represents very important. After<br />

answer sheet gathering, first the stability of given answers was calculated by SPSS. It was 91.2 percent<br />

which indicates that IT expert’s answers were at minimum deviation.<br />

Second, the mean value of each criterion and domains were calculated. The cut off value of 3.0 within<br />

values was assumed and those criteria which their mean values are greater or equal 3.0 were identified<br />

and ignored. The 3.0 value is almost moderate importance of each criterion. Based on this assumption,<br />

just one of the sub criteria has been eliminated and excluded from the model. This criterion was<br />

“improvement in organizational communication” from the “effects” criteria. This condition shows that all<br />

these criteria are important and reduction process in the previous step was successful to capture a<br />

minimal set of attributes for IT investment evaluation.<br />

Third, as a consequence based on mean values the entire criterions were ranked. Table3 shows the<br />

gathered results. According to this table, security criteria are the most important and “Improvements<br />

internal and external organizational culture” criterion has less importance within all criteria.<br />

After that, the overall weighting of criteria and domains were calculated with dividing each mean value by<br />

the total sum of its related criteria and sub criteria mean values. This is called prioritizing. By prioritizing<br />

them, the level of each criterion would be determined in its criteria and in an analysis hierarchy model<br />

they would be shown at decreasing level. Figure2 illustrates the hierarchy model based on the<br />

decreasing prioritization.<br />

According to mentioned model, strategic domain has more impact on the investment final decision<br />

making than operational domain. Indeed, nowadays it is the fact organization strategy and IT investment<br />

alignments with strategies are so important for chief managers.<br />

329


Table 3: Evaluation criteria and statistical results<br />

Domain<br />

Strategi<br />

c<br />

0.516<br />

Operati<br />

onal<br />

and<br />

Tactical<br />

0.484<br />

Criteria<br />

Strategic Alignment<br />

0.287<br />

Competitive<br />

Advantages<br />

0.167<br />

Organizational<br />

productivity<br />

0.267<br />

Market Assessment<br />

0.148<br />

Compliance<br />

0.131<br />

Quality<br />

0.217<br />

Security<br />

0.247<br />

Flexibility<br />

and<br />

compatibility<br />

0.133<br />

Cost<br />

0.183<br />

Satisfaction<br />

0.117<br />

Effects<br />

0.103<br />

Shirin Nasher et al.<br />

Sub Criteria<br />

Impact on achieving IT strategic<br />

objectives<br />

Impact on aligning IT with organization<br />

strategy<br />

Organization governance<br />

Improved efficiency of current business<br />

process<br />

Improved efficiency of changed business<br />

process<br />

Improved efficiency of new business<br />

process<br />

Customers<br />

Partners<br />

Suppliers<br />

Competitors<br />

New incoming<br />

Increase organizational productivity<br />

The number and credit of previous<br />

customers<br />

Competitor opinion<br />

Define market<br />

Market anticipating<br />

Internal standards and requirements<br />

External standards and requirements<br />

Response Time<br />

Turnaround Time<br />

Error Rates<br />

Program Code Quality<br />

Quality Assurance<br />

Customization of Outputs<br />

Outputs Format<br />

Comparability of output produced<br />

Easy output format<br />

Security( Confidentiality, Integrity,<br />

Availability)<br />

Flexibility against technological change<br />

compatibility<br />

Ability to customize<br />

Update Feature<br />

Facilitate changes in the code<br />

Buy<br />

Implementation<br />

Human Resources Training<br />

Maintenance<br />

Time<br />

Internal satisfaction<br />

External satisfaction<br />

Improvements internal and external<br />

organizational culture<br />

Importance<br />

Average<br />

4.5<br />

4.65<br />

4.14<br />

4.1<br />

3.72<br />

3.7<br />

4.55<br />

4.35<br />

3.4<br />

4.3<br />

3.1<br />

4.63<br />

3.91<br />

3.55<br />

3.85<br />

3.83<br />

3.11<br />

3.2<br />

4.23<br />

3.90<br />

4.31<br />

3.75<br />

4.05<br />

3.73<br />

3.89<br />

3.18<br />

4.17<br />

4.85<br />

4.25<br />

4.4<br />

4.2<br />

3.68<br />

3.35<br />

4.33<br />

4.6<br />

3.8<br />

4.15<br />

3.95<br />

3.65<br />

3.88<br />

3<br />

Rank<br />

6<br />

2<br />

17<br />

18<br />

30<br />

31<br />

5<br />

8<br />

35<br />

11<br />

40<br />

3<br />

21<br />

34<br />

25<br />

26<br />

39<br />

37<br />

13<br />

22<br />

10<br />

28<br />

19<br />

29<br />

23<br />

38<br />

15<br />

1<br />

12<br />

7<br />

14<br />

32<br />

36<br />

9<br />

4<br />

27<br />

16<br />

20<br />

33<br />

24<br />

41<br />

Weight<br />

0.181<br />

0.184<br />

0.166<br />

0.165<br />

0.150<br />

0.149<br />

0.231<br />

0.221<br />

0.172<br />

0.218<br />

0.157<br />

1<br />

0.258<br />

0.234<br />

0.254<br />

0.253<br />

0.493<br />

0.507<br />

0.120<br />

0.111<br />

0.122<br />

0.107<br />

0.115<br />

0.106<br />

0.110<br />

0.090<br />

0.118<br />

1<br />

0.214<br />

0.221<br />

0.211<br />

0.185<br />

0.168<br />

0.208<br />

0.221<br />

0.182<br />

0.199<br />

0.190<br />

0.483<br />

0.516<br />

The security criteria are the most important criteria and the impact of it on the choosing IT project is much<br />

more. Also, the prioritizing sub criteria can help to have a scale to understand the impact of them on the<br />

investment decision. Note that, the sum of sub criteria weights is one and it shows their overall impacts<br />

are considered on the final investment decision making.<br />

330<br />

1


Figure 2: Analytic hierarchy decision making model<br />

3.4 Proposing model<br />

Shirin Nasher et al.<br />

Since the model wants to illustrate the criteria importance level on the final decision making process,<br />

another hierarchy model was proposed. In this model, two domains were determined on the top of the<br />

model. This model is comprised hierarchical level based on five unit intervals that the mean value of each<br />

criterion determine its own level. By comparing these values, they have been assigned to their<br />

appropriate level as shown in Figure3.<br />

There was an unanimous consensus that this proposed model has improved the decision process in<br />

order to have consistent and complete criteria based on different tangible and intangible factors for IT<br />

investment evaluation. Note that this proposed model is general model and independent from the specific<br />

IT investment to aid investment decision making process.<br />

4. Conclusion<br />

Adoption of the new information technology requires investment justification. Also, IT investment decision<br />

making is based on evaluating tangible and intangible criteria. But the intangible criteria are complex to<br />

evaluate and therefore this action may take a long time and efforts. So many managers prefer to neglect<br />

them and just focus on the financial factors. They think that IT investment affects on their organization<br />

and enhances productivity efficiency in situ. But indeed return on IT investment has a potential latency to<br />

create value just because of these intangible factors. So it shows the importance of these criteria to have<br />

an appropriate investment evaluation. The existence of the pre-defined criteria hierarchy model based on<br />

the consensus of many specialist, can structured the decision making process systematically. Hence this<br />

article proposed criteria list and model can significantly decrease times and efforts for managers and<br />

organizations.<br />

So in this paper, we first determine these tangible and intangible criteria and then categorize them in<br />

different criteria and sub criteria. As a consequence, we give importance rank to them according to the<br />

experiment of chief managers in Iran. Based on these ranking, at last we proposed the hierarchy model in<br />

331


Shirin Nasher et al.<br />

order to consider all these tangible and intangible criteria in IT investment evaluation. Note that since<br />

there are many sub criteria and they may be changed from one IT evaluation to another, we try to<br />

consider general criteria to create a generic model. So this proposed model is applicable and can be<br />

easily expanded by aggregating new sub criteria to be customized for different IT investment evaluations.<br />

Figure 3: Final hierarchy decision making model<br />

References<br />

Borenstein, D. and Baptista Betencourt, P. R. (2005) “A Multi-criteria Model for the Justification of IT Investments”,<br />

INFOR Journal, Vol. 43, No. 1, Feb, pp. 1-21.<br />

Chen, T., Zhang, J. and Lai, K. (2009) “An Integrated Real Options Evaluating Model for Information Technology<br />

Projects under Multiple Risks”, International Journal of Project Management, Vol. 27, Issue 8, pp.776–786.<br />

Epstein, M. and Rejc, A. (2004-2005) “Measuring the payoffs of IT investments”, CMA Management Journal, Vol. 78,<br />

No. 8, pp. 20-25.<br />

Goh, K.H. and Kauffman, R.J. (2005) “Towards a Theory of Value Latency for IT Investments”, Proceedings of the<br />

38th Hawaii International <strong>Conference</strong> on System Sciences, pp. 1-9.<br />

Goldstein, Ph., Katz, R.N. and Olson, M. (2003) Understanding the Value of IT, EDUCAUSE Quarterly magazine,<br />

Number 3, pp. 14-18.<br />

Henderson, J. and Venkatraman, N. (1990) Strategic Alignment: A model for Organizational transformation Via<br />

information technology,<br />

Joseph Wen, H., Yen, D.C. and Lin, B. (1998) “Methods for measuring information technology investment payoff”,<br />

Human Systems Management Journal, Vol. 17, No. 2, January, pp. 145–153.<br />

Kraemer, K. and Dedrick. J. (1994) “Payoffs from Investment in Information Technology: Lessons from the Asia-<br />

Pacific Region”, World Development Journal, Vol. 22, No. 12, pp. 1921-1931.<br />

Mahmood, M.A., Szewczak, E.J. (1999) Measuring Information Technology Investment Payoff: Contemporary<br />

Approaches, Idea Group Publishing, London.<br />

Parker, M.M. and Benson, R.J. (1988) Information Economics: linking business performance to information<br />

technology, Prentice-Hall.<br />

Renkema, Th. J.W. and Berghoutb, E.W. (1997) “Methodologies for information systems investment evaluation at the<br />

proposal stage: a comparative review”, Information and Software Technology Journal, Vol. 39, Issue 1, pp. 1-<br />

13.<br />

Schniederjans, M. and Hamaker, J. (2003) “A new strategic information technology investment model”, Management<br />

Decision Journal, Vol. 41, No. 1, pp. 8-17.<br />

Schniederjans, M.J., Hamaker, J.L. and Schniederjans, A.M. (2004) Information Technology Investment Decision-<br />

Making Methodology, World Scientific Publishing Co., London.<br />

Tuten, P. M. (2009) “A Model for the Evaluation of IS/IT Investments”, A dissertation submitted in partial fulfillment of<br />

the requirements for the degree of Doctor of Philosophy in Information Systems, Graduate School of Computer<br />

and Information Sciences Nova Southeastern University, USA.<br />

Wu, L.C. and Ong, C.S. (2008) “Management of information technology investment: A framework based on a Real<br />

Options and Mean–Variance theory perspective”, Technovation Journal, Vol. 28, Issue 3, March, pp. 122-134.<br />

332


Critical Factors in the use of Mobile Learning by “Digital<br />

Natives” on Portuguese Teaching<br />

Mário Carrilho Negas and Paulo Ramos<br />

Universidade Aberta, Portugal<br />

mnegas@univ-ab.pt<br />

rcjp@sapo.pt<br />

Abstract: The purpose of this paper is to discuss the mLearning as a useable resource in formal education (in a<br />

classroom context), in order to improve the learning experience of young people. It is intended to enumerate the<br />

possible success factors that can contribute to the mLearning and to show that, when properly integrated into school<br />

curricula, they could be a valid learning experience for youth. Young people are referred to as the "digital natives"<br />

because of their constant use of digital resources in their day-to-day. Some factors of significant importance are<br />

identified, namely the fact that young people see the mLearning as a valid learning tool, and that they accept its use<br />

in teaching, provided it is not to replicate contents taught in the classroom. This article highlights the important role<br />

that mobile learning can have on academic life, and the factors which underpin its success, such as the widespread<br />

use of mobile devices by young people. The "Digital Natives" use mobile technology with great frequency, particularly<br />

to keep in touch with family and social networks as they see the technology as an essential part of their lives. Mobile<br />

technology solutions may help transforming the way to create, teach and learn; in this sense, mobile phones may<br />

have an important role to play in tomorrow’s school.<br />

Keywords: mobile learning, eLearning, moving learning capabilities; mobile devices; wireless network; digital<br />

natives; teaching<br />

1. Introduction<br />

The constant evolution of the technologies, as it is the case of the mobile technologies and particularly<br />

mobile phones, has made possible a significantly wide set of solutions that help to change the way we<br />

create, teach and learn. The ubiquity of mobile technology in our day-to-day (Green et al, 2001, Katz and<br />

Aakhus, 2002; Levinson, 2004), and this fact has contributed to the emergence of a new educational<br />

paradigm, the mobile learning (mLearning). For Prensky (2001) and Kukulska-Hulme (2005) the concept<br />

of mobile learning was an evolving subject that is no longer associated exclusively with the<br />

characteristics of portability and mobility of the equipment. On the contrary, it must be understood<br />

nowadays as a concept associated with the mobility of the user.<br />

The ancestral human search for knowledge has led to the search for newer and more sophisticated ways<br />

to store, process and transmit information, leading to the evolution of tools which enable that search.<br />

Palfrey, J. and Gasser, U. (2008) state that although several decades of constant technology evolution<br />

and the resulting saturation of the digital technologies in some cultures, there wasn’t yet a generation that<br />

has lived exclusively in the age of the digital technology.<br />

In 2007, the Portuguese government launched a major public initiative – especially given its extent – in<br />

order to distribute laptops with Internet access to children and youth. This helped creating the right<br />

conditions for the use of information and communication technologies (ICT) in education since ICT<br />

learning occurs at a younger age.<br />

In the 60´s, Herbert Marshall McLuhan predicted that the planet would become a teaching room and our<br />

virtual address.<br />

Learning environments have acquired new characteristics; they are more open, more interactive and<br />

flexible. These characteristics may allow combining different learning styles. The use of mLearning in this<br />

context is an excellent opportunity for affirmation of this tool in the context of teaching-learning<br />

environment.<br />

The evolution of mobile devices may enable – if used properly by education agents – the access to new<br />

knowledge and the review of previously acquired knowledge.<br />

333


Mário Carrilho Negas and Paulo Ramos<br />

2. The "Digital Natives" and the teaching-learning<br />

Palfrey and Gasser (2008) stated that the first generation of "Digital Natives" is the current youths. In fact,<br />

this generation is the first one that was born and has lived in the digital age since birth. According to the<br />

authors, this will contribute to a major overhaul in the way the world is organized.<br />

The way young people perceive the world will have implications on issues of political, cultural and family<br />

life, which will be changed forever. Students today are questioning the current practice and breaking with<br />

the so-called traditional guidelines (Boyle, 2008), questioning some of the practices of the 20th Century.<br />

In comparison to the "Digital Immigrants" ("Digital Immigrants"), the “Digital Natives” have some<br />

differences in speech, social interaction, and are fluent in using new ICT. It is possible to identify three<br />

generational groups: Baby Boomers (1946-1964), the "Generation X" (1965-1977) and the "Digital<br />

Natives" (1978-1994). The latter, in contrast to previous generations, tend to use information they gather<br />

from various online sources.<br />

The use of mobile technologies by young people tends to overcrowding. It is relevant to know to what<br />

extent the use of mLearning can contribute as a motivational factor in the process of teaching and<br />

learning of young people. In this sense the institutions and teachers are aware of this reality: how young<br />

people react to this technology, how they adopt it, and if the educational programs of schools enable the<br />

use of these tools.<br />

The discussion of new technologies in education is a constant in the adoption of new methodologies;<br />

however, to digital natives this is not an issue, because younger generations begin academic life with no<br />

doubts or fears (Tori, 2010).<br />

Young people tend to identify the use of mLearning as a possible tool for learning. They also tend to<br />

accept the use of mLearning in their courses, as long as that is not only a repetition of educational<br />

content taught in the classroom, and preferably not just an isolated initiative, but rather a practice<br />

supported by several teachers. The learning content to be provided must be diversified, classified<br />

according to the subjects being taught, and adapted to the possibility of using mobile phones at low cost<br />

and to the prevalence of information in audio and video formats instead of text. The fact that the "Digital<br />

Natives" use technology from an early age eases the process of constant adaptation to technological<br />

change that generally occurs. It's with recognized ability that "Digital Natives" the hardware to their goals<br />

of recreation, using the online environment more than any other preceding generation.<br />

According to Tori (2010), the "Digital Immigrants" and "Digital Natives" have points of convergence. Even<br />

though that may be true, the author mentions that the "Digital Immigrants" contributed significantly to the<br />

“Digital Natives” empathy with ICT.<br />

Despite having been taught by the “Digital Immigrants”, the "Digital Natives" began to adopt new<br />

practices in the use of ICT. Generally, for “Digital Natives”, a day’s time is divided between online and<br />

offline interaction, "nick name" and name, virtual friendships on social networks and real life friendships.<br />

Even though this quick change in reality may be confusing for those who are not "Digital Natives", it is<br />

something intrinsic to the present reality for those who belong to that group. When entering the job<br />

market, this generation carries the practices associated with the digital world. This young generation has<br />

a significant set of distinctive skills in comparison with the current generation, the Generation X.<br />

Virtual social networks are far superior in dimension than the neighborhood, academic or working social<br />

networks, and may also be viable sources of information to be used in collaborative learning<br />

environments: they cover very different subjects, are very heterogeneous and culturally diverse, which<br />

allows the obtainment of diversified and relevant information. However it is also questioned if their<br />

contribution as an educational resource in teaching-learning process may constitute a source of<br />

distraction and waste of time, taking into account the objectives of educational programs (Economist<br />

debates: Social networking ", 2008).<br />

Boyle, L. and Stanford, J. (2008) sustain that “digital natives” are native speakers of technology, fluent in<br />

the digital language of computers, video games and the Internet. They also state that the students, as<br />

digital natives, will continue to evolve and change so rapidly that they will have, in general, a much better<br />

idea of what the future will bring. Since the natives are already busy adopting new systems for<br />

334


Mário Carrilho Negas and Paulo Ramos<br />

communications (instant messaging), sharing (blogs), buying and selling (eBay), exchanging (peer-topeer<br />

technology), creating (Flash), meeting (3D worlds), collecting (downloads), coordinating (wikis),<br />

evaluating (reputation systems), searching (Google), analyzing (SETI), reporting (camera phones),<br />

programming (modding), socializing (chat rooms) and even learning (Web surfing), they have a better<br />

understanding of their surroundings than previous generations.<br />

3. The mobile learning in distant teaching<br />

Pachl et al. (2009) examined the characteristics identified as important for mobile devices in order to<br />

make them so attractive to users in the education field. Portability is one of those factors, as the use of<br />

portable equipment in various life situations became socially acceptable, and naturally they have a vast<br />

potential for providing access to culture, content and features, such as the possibility to record<br />

information in audio, video or text. Another very interesting factor is the number of applications or<br />

services that are can be accessed through a mobile device, some of which being accessible also through<br />

the Internet.<br />

The mobile learning (mLearning) is presented as one of the possible tools of distance learning,<br />

specifically in eLearning. Pachl et al. (2010) state that learning with the use of mobile devices is an<br />

emerging and growing field of research due to the increasing use of mobile equipment in schools,<br />

universities and workplaces. The use of mobile networks is gaining greater importance in unconventional<br />

places. These and other factors have contributed to the growing interest of researchers, particularly in the<br />

area of educational research.<br />

Mobile Learning is an educational perspective, a new dimension in education that could meet immediate<br />

learning needs with great flexibility and interactivity.<br />

Undoubtedly, one of the great advantages of mobile devices is their accessibility, which allows constant<br />

contact with other users (either because the device can always go with the user, or because of the instant<br />

access it allows to the communication network or information requested).<br />

Being easily carried and used in periods of dead time (such as travel or waiting periods), these<br />

applications become an excellent way to access learning content and information in general.<br />

The lifelong learning is increasingly seen as a constant practice. Therefore, there is a great expectation<br />

about whether mobile technologies have the potential and feasibility to be used in lifelong learning<br />

systems in a large scale.<br />

4. The role of mobile technologies in the increase of the sustainability of<br />

distance teaching<br />

When it was created, the cell phone was meant to keep people in touch. Now it has become a true all-inone<br />

device. It enables us to connect to the Internet, e-mail, listening to music, watching TV or even<br />

checking our bank accounts online. Wireless mobile networks have suffered quick qualitative changes<br />

throughout times, since the first analogic generation (1G) became digital (2G). It became a massive<br />

service by the third generation (3G), which integrated wireless digital transactions and services that were<br />

traditionally available in landlines (Harte, Levine & Kikta, 2002), and included better control of information<br />

packages, connections speeds of up to 2 Mbps, multiple radio channels, among other features.<br />

The Long Term Evolution technology (LTE), commonly named 4G, will be the future tendency to fulfill the<br />

requirements of the Digital Natives, and will enable faster connections, with maximum debits of 100/150<br />

Mb on downloads and 50 Mb on uploads. LTE presents a greater efficiency in using the radioelectric<br />

space, and a larger mobile networking capacity than the current 3G, which presents maximum download<br />

speeds of 7,2 Mb and upload speeds of 384 Mb.<br />

It is expected for LTE to be available in 2011, progressively in FDD mode on Europe. Greater speed and<br />

efficiency will also serve as a countermeasure against the expected increase in data traffic. Teachers<br />

may include the technologies available in their teaching methodologies, so that they can provide an<br />

enriching experience to their students (Guy, 2009).<br />

Logistically speaking, LTE may sustain the academic experiences, since it keeps up several bandwidths,<br />

which has special importance in a time when mobile broadband has been growing exponentially. LTE will<br />

also enable the opening of new possibilities, such as services and functionalities that require high<br />

335


Mário Carrilho Negas and Paulo Ramos<br />

transmission debits, like the advanced diffusion of television via internet, online games and web<br />

conferences, whether there is a high or low mobility associated.<br />

The hardware, software and networking quality are critical factors. A recent study (Guy, 2009) by<br />

Schmidt, Hodge, Dickerson, and Ellis (2008) reveals that students often enroll in online courses due to<br />

the schedule flexibility, the information available and its formats (audio, video, among others), instead of<br />

conventional classes, even when the universities and academic campus are geographically close.<br />

As there is an increase in the amount and virtual space of the information that is transacted online,<br />

devices and networks need to adjust to the Digital Natives’ impatience – who want it all on the spot<br />

without waiting. The development of new technologies (hardware, software, telecommunications,<br />

networks and Internet) happens at a frantic rate and the lifestyle of the Digital Natives in their schools and<br />

surrounding social environments has also been acquiring an increasingly accelerated pace. Mobile<br />

technology is more present in our day-to-day academic or professional life (Guy, 2009) with the huge<br />

advantage of not needing a conventional physical space, electricity connection of cable connection to the<br />

Internet. The sales of informatics equipment – notably laptops – have ascended to the 88,3 millions on<br />

the first trimester of the current year worldwide, according to the news, which means that a computer<br />

went from being a superfluous item to become a first necessity tool.<br />

5. Opportunities and challenges of mobile learning in Portugal<br />

Studies about the young users of the technologies of communication and information began in 2006<br />

(Cardoso et al., 2007), with the study entitled “E-Generation - The Uses of Media for Children and Youth<br />

in Portugal”. The methodology used by that study was the application of an online inquiry to young<br />

Internet users. This inquiry was completed 1377 times during the month it was online, from which 1353<br />

entries were considered valid, allowing a characterization of young Portuguese Internet users and their<br />

relationship with technologies and the means of communication. In 2008 this study was conducted again<br />

to deepen the knowledge and the comprehension of the data previously obtained in 2006. The study<br />

highlighted several behavioural characteristics common to children and youths, namely the fact that the<br />

socialization space is now online, in chats, blogs, and social networks. While trying to strengthen their<br />

social networks, cell phones have played an important role as a mean of socialization, but also as an<br />

agent of personal education. According to these studies, most of the young people inquired began using<br />

the Internet at an early age, between 10 and 12 years old, and it is expectable that they tend to do it<br />

increasingly early (Cardoso, et al, 2007, 395). Therefore, young people develop more and more early a<br />

mentality focused on the use of the technologies of information and communication, as well as a range of<br />

skills centered on multitasking (McMahon and Pospisil, 2005, p. 421). The technology and the means that<br />

young people use to communicate during their school years tend to influence the way they relate to the<br />

information and consequently the method by which they prefer to learn. Several authors attribute to these<br />

young people the ability to stay in contact through SMS and watch television, whilst simultaneously<br />

playing computer games and listening to music (Frand, 2000, p. 18; Oblinger, 2003; Rickard and<br />

Oblinger, 2003). The study entitled E-Generation, 2006 (Cardoso et al., 2007) points out that 96,6% of<br />

people between eight and eighteen years old (with a sampling universe of 1353 inquiries) has a cell<br />

phone, which reveals its high utilization rate. In Portugal, the average age at which the inquired people<br />

acquired a cell phone is of 11,8 years old. It is, without a doubt, important for Portugal, but only the near<br />

future will reveal if this fact has contributed to the enhancement of young people’s skills in a more<br />

generalized way. A study conducted by Pew internet & American Life Project reveals that 55% of the<br />

American young Internet users between the age of 12 and 17 are registered on social networks. (Lenhart<br />

e Madden, 2007). Valkenburg, Peter and Schouten (2006) also refer that in Holland about a quarter of<br />

teenagers is currently a member of one or more social networking websites. By analyzing the Top 5 of<br />

the most researched expressions on the Internet in Portugal in 2003, 2005 and 2007, we find, in<br />

decreasing order of hits, in 2003: Portugal; Download; Lisbon; Free; Oporto; in 2005: Sex; Emule; Google<br />

earth Sapo; Benfica; in 2007: Youtube; HI5; Gmail; Wikipedia; Games. It is possible to verify that the<br />

most researched expressions on the Internet in Portugal in 2007 include 5 communication-related<br />

expressions. A study entitled e-Generation 2008 (Cardoso et al., 2008), focused specifically on youth,<br />

from 8 to 18 years old, (with the sampling universe composed of 489 validated inquiries), yielded the<br />

following results: 78,8% of the people who were inquired use e-mail; 70,5% use the IM network; 63,3%<br />

use social networking sites; 43,1% send SMS through the Internet, and 12,4% of the inquired make<br />

phone calls through the Internet. Although the Internet is more and more used by young people for<br />

communication purposes, the mobile phone remains the most used equipment, followed by the TV, and<br />

only then the Internet. In voice communication the mobile phone remains the most widely used<br />

equipment by young people to maintain their social contacts, and it is also the most used to keep them in<br />

336


Mário Carrilho Negas and Paulo Ramos<br />

touch with family members, and as a way to organize tasks and daily meetings (using SMS and phone<br />

calls).<br />

6. The "e-escolinha" and "e-escola" as a driver of mobile learning in Portugal<br />

Guy (2010) affirms in his book that the overall trends in mobile learning technology are becoming clear,<br />

and that devices will continue to have greater processing capacity, greater storage capacity, faster<br />

network access, and will be carried and accessed more and more frequently due to the students’<br />

necessity for research capabilities. Although Portugal is a developing country on many levels, there have<br />

been many physical, logistical and behavioural advances in what refers to the ICT, allowing the country to<br />

minimize or even dissolve the distance to some other more developed countries on this subject. With this<br />

goal in mind, it was created a program to distribute laptops “Classmate PC” at reduced cost for school<br />

children. These laptops were called Magalhães (Magellan) and were manufactured in Portugal with an<br />

Intel processer.<br />

The Program was created to promote the access to the Information Society and to foster info-inclusion,<br />

through the availability of laptops and internet broadband connections, and it has registered an adhesion<br />

of 1 million people in the first two years.<br />

The Program is divided into three parts: e-schools for children from the 5th to the 12th year, and small<br />

schools for children in the first year of school. The Technological Plan for Education has provided<br />

students (especially the most financially needy) with the opportunity to possess a computer, allowing<br />

them to heuristic development, autonomy, and to serve as a complement to their personal development.<br />

This initiative was launched in July 2007 in partnership with the major Portuguese telecommunication<br />

operators, such as Optimus, Vodafone and TMN, including students from the 1st to the 12th grade,<br />

teachers and adult students as well. Several protocols were also signed with software and hardware<br />

designers, to enable the creation of a richer educational experience.<br />

Consequently, the Program was recognized by its effort and awarded on November 2007 with the “Best<br />

<strong>European</strong> Project Award” by Toshiba Europe, represented by the president of the company Noriaki<br />

Hasimoto. In January 2009, in London, Microsoft indicated the Program e-escolas as an example to be<br />

followed by other governments.<br />

The final results of these plans can be seen as an improvement in the students performance and facilitate<br />

the access to reasonable hardware and Internet access. Overall, this project intends to achieve several<br />

objectives, namely: giving opportunities to children and youth from the 5th to the 12th year, younger<br />

students aged 6-10, and developing the technology and the software industries. It should be noted that,<br />

in recent years, Portugal has taken measures to develop these industries, such as the case of mobile<br />

software. Several measures were developed to create a larger offer of Internet services, according to the<br />

2010 annual report from Bareme Internet. Besides this increase, the Portuguese municipal facilities often<br />

include free internet services in libraries and other public areas. In what concerns the private sector, there<br />

is an emergence of internet cafés in which Wi-Fi services offer convenient and inexpensive access to the<br />

internet. There are also conditions for the students to reduce mobile internet monthly payments.<br />

7. Use of mobile learning on Portuguese teaching<br />

Distance learning or distance education (EaD) in Portugal has existed for several years (Cardoso and<br />

Machado, 2001). Currently, resorting to the ICT, the use of the ELearning has become a teaching tool<br />

that can be adapted to the agitated daily life of our time, and that represents a solution for students that<br />

are geographically distant from education institutions or have mobility problems.<br />

According to Leal and Amaral (2004), with the dawn of the 90’s appeared different teaching-learning<br />

practices, supported by the Mobile Learning technologies, that allowed the online participation of each<br />

student/teacher in either synchronous or asynchronous mode. Since then, the Internet utilization rate has<br />

grown exponentially (in Tek, 13 de Outubro de 2010), as well as the number of users. In fourteen years,<br />

the penetration rate grew by 920 per cent, an average of 20 per cent a year. In the present year, 57,1 %<br />

(4 749 million) of the Portuguese periodically access the Internet. Judging by these numbers, it is<br />

possible to perceive the importance of this technology to research information.<br />

337


Mário Carrilho Negas and Paulo Ramos<br />

Figure 1: Number of internet users in Portugal (source: Marktest - Bareme Internet (2010))<br />

Young internauts with ages from 15 to 17 are the group more often connected to the Internet (99,2%) (<br />

Marktest - Bareme Internet, 2010).<br />

The project “mLearning – The role of mobile learning in <strong>European</strong> Education” developed by TecMinho,<br />

Minho University, in association with other institutions from northern Europe, is a reference in the<br />

implementation and development of mobile learning. This project is managed by Ericsson (from Ireland)<br />

and is inserted in the Program Sócrates of the <strong>European</strong> Union, which has been financing mLearning<br />

activities in Portugal. The main outcome of this program will be:<br />

The development of a document that may serve as guideline for politics about the role of mLearning.<br />

The analysis of the initiatives and the role that mLearning currently plays on the 25 members of the<br />

EU;<br />

The study of good practices and current projects concerning mLearning worldwide.<br />

Analysis of the pedagogic aspects of mLearning;<br />

Adaptation of existing learning contents to mLearning;<br />

Development of several formation courses in mLearning by the partner entities involved in the<br />

project;<br />

A scientific analysis of the advantages of mLearning for handicapped people and of the strategies for<br />

presenting learning contents to these students.<br />

According to Junior and Coutinho (2008), of the 1166 individuals who answered the electronic enquiry,<br />

88% claimed to use a mobile phone and 11% stated they also had a PDS or a Smartphone besides their<br />

mobile phone. However, when enquired about mLearning, more than a half of the people didn’t know this<br />

concept of learning, and only 39% said that they recognized the concept applied to teaching and learning<br />

activities. However, 25% had used mobile phone, 15% the tablet computer, 7% the pocket PC and 5%<br />

the Palm or PDA, and the great majority, 48%, revealed they never used any kind of device.<br />

As stated by the study “E-Generation - The Uses of Media for Children and Youth in Portugal”, only a<br />

small minority of young people (5,3%) possesses a laptop and also a minority (6,9%) uses internet<br />

through wireless devices, like the the mobile phone or a laptop (Cardoso et al., 2007).<br />

In spite of being a part of young people’s daily lives, even in the classroom, the eventual potential of<br />

mobile technologies has yet to be explored in eLearning environment.<br />

8. Conclusions<br />

It may be said that the cell phone is a piece of technology which is relatively inexpensive to the average<br />

citizen when compared with other mobile devices. Additionally it is easy to use and is very widespread<br />

338


Mário Carrilho Negas and Paulo Ramos<br />

nowadays. Young people constitute a large group of cell phone users, and they use it in several activities<br />

during the day, for various social and family contacts and for schedule management. In this sense the<br />

utilization of mobile learning as an aid for teaching-learning is a strategy that is gaining supporters as<br />

mobile devices evolve. However, it is known that the “Digital Native” community doesn’t easily accept<br />

mobile learning as a learning tool if it only duplicates the contents that are taught in the classroom. As far<br />

as “Digital Natives” are concerned, mobile learning should be a complement to the classroom learning, in<br />

order to make the learning experience more interactive. Mobile learning allows the enrichment of the<br />

pedagogic offer in a way that fits the students experience outside the classroom, making it faster, easier<br />

and more flexible to access and share information. Even so, there are some difficulties to implement the<br />

mobile learning, namely the fact that the young public is by nature eager to know new things: through the<br />

utilization of the resources already available in the Web 2.0, they acquire innovative and enriching<br />

experiences that may drive them to place their expectations higher than it is possible for teachers and<br />

education institutes to keep up with. If it is accepted that “Digital Natives” have access to technology and<br />

are acquainted with mobile learning, using it constantly, it must be considered, on the other hand, the<br />

teachers’ necessity to also acquire the necessary competences to enable them to provide systematic<br />

answers to the various challenges, namely the creativity, cooperation, communication and involvement<br />

that the mLearning requires (Oblinger, G. (2003). According to Facer, Faux, and McFarlane (2005) the<br />

deficit of technical preparation and the lack of confidence and practice of mLearning that is common to<br />

teachers can have a negative effect on the involvement and pedagogic use of this resource by the<br />

“Digital Natives”. In this sense, Sharples, Taylor and Vavoula (2005, p. 4) identify some of the pedagogic<br />

difficulties concerning mLearning, namely the fact that the pedagogic practices in a classroom are<br />

different and it is intended to cover both formal and informal components of teaching and learning. For<br />

Bryant (2006) and Thomas (2005), the involvement of the “Digital Natives” in the preparation of<br />

pedagogic contents in a cooperative way, “user-led-education”, may be stimulating for the students. Over<br />

the past few years, Portugal has been undertaking efforts to broaden the use of the ICT by young people.<br />

However, the teacher’s instruction and the allocation of funds to support the development and application<br />

of the mLearning on Portuguese schools has been very insufficient. The fact that thousands of laptops<br />

“Classmate PC” have been distributed to children and youth from the 5 th to the 12th grade and also in<br />

elementary schools for children in the first grade, combined with the decreasing costs of mobile<br />

communications, constitutes an important measure, if associated with the high penetration rate of mobile<br />

devices near young people.<br />

References<br />

Ally, M. (2009). Mobile Learning Transforming the Delivery of Education and Training, AU Press, Athabasca<br />

University.<br />

ANACOM (2006). Serviços de Comunicações Electrónicas Móveis, 2.º Trimestre de 2006. (Retrieved November 15,<br />

2010, from: http://www.anacom.pt/streaming/smrp_2t06.pdf)<br />

ANACOM (2009). Serviços de Comunicações Electrónicas Móveis, 2.º Trimestre de 2009. (Retrieved November 15,<br />

2010, from: http://www.anacom.pt/render.jsp?contentId=970592)<br />

Bareme Internet. (Retrieved November 10, 2010, from:<br />

http://www.marktest.pt/marktest/default.asp?c=1210&n=2085)<br />

Boyle, L. and Standford, J. (2008). Digital Natives. Morlacchi Editore, Perugia.<br />

Bryant, T. (2006). Social software in academia. EDUCAUSE Quarterly, 2.( Retrieved November 22, 2010, from:<br />

https://www.educause.edu/ir/library/pdf/eqm0627.pdf)<br />

Cardoso, E., and Machado, A. (2001). Problemática da Adopção de Plataformas eLearning nas Instituições de<br />

Ensino Superior - a fase de iniciação. Associação Portuguesa de Sistemas de Informação, nº 15, 49.<br />

Cardoso, G., Espanha, R. and Lapa, T. (2007). E-Generation: Os Usos de Media pelas Crianças e Jovens em<br />

Portugal. (Retrieved September 27, 2010, from:<br />

http://www.cies.iscte.pt/destaques/documents/CriancaseJovens_AsuarelacaocomasTecnologiaseosMeiosdeCo<br />

municacao.pdf)<br />

Economist debates: Social networking (2008). Economist.com. (Retrieved September 25, 2010, from<br />

http://www.economist.com/debate/overview/123).<br />

Facer, K., Faux, F., and McFarlane, A. (2005). Challenges and opportunities: Making mobile learning a reality in<br />

schools. Proceedings of mLearn 2005. (Retrieved November 5, 2010, from<br />

http://www.mlearn.org.za/CD/papers/Facer%20-%20Faux%20-%20McFarlane.pdf)<br />

Frand, J. (2000). The information age mindset: Changes in students and implications for higher education.<br />

EDUCAUSE Review, 35(5), pp.15-24.<br />

Green, N., Harper, R., Murtah, G., and Cooper, G. (2001). Configuring the Mobile User: Sociological and Industry<br />

Views. In Personal and Ubiquitous Computing (pp. 146-156). London: Springer.<br />

Guy, R. (2009). The Evolution of Mobile Teaching and Learning. Informing Science Press, California, pp.57- 58.<br />

Guy, R. (2010). Mobile Learning: Pilot Projects and Initiatives. Informing Science Press. California, pp.1-7.<br />

Harte, L., Levine, R. and Kikta, R. (2002). 3G Wireless Demystified. McGraw-Hill. NY.<br />

339


Mário Carrilho Negas and Paulo Ramos<br />

Junior, J. and Coutinho, C. (2008). The use of Mobile Technologies in Higher Education in Portugal: an exploratory<br />

survey. (Retrieved November 30, 2010<br />

(http://repositorium.sdum.uminho.pt/bitstream/1822/8468/1/Elearn_Tec_Moveis.pdf)<br />

Katz, J. and Aakhus, M. (2002). Perpetual Contact: Mobile Communication, Private Talk, Public Performance.<br />

Cambridge: Cambridge University Press.<br />

Kukulska-Hulme, A. (2005). Introduction. In Kukulska-Hulme, A. and Traxler, J. (eds).. Mobile Learning: A Handbook<br />

for Educators and Trainers. London: Routledge, pp.1-6.<br />

Leal, D. and Amaral, L. (2004). Do ensino em sala ao eLearning. Guimarães: Universidade do Minho.<br />

Lenhart, A. and Madden, M. (2007). Social Networking Websites and Teens: An Overview. Pew Internet and<br />

American Life Project. (Retrieved November 10, 2010.<br />

http://www.pewinternet.org/PPF/r/198/report_display.asp)<br />

Levinson, P. (2004). Cellphone: The story of the world's most mobile medium and how it has transformed everything.<br />

Nova Iorque: Palgrave Mcmillan.<br />

McLuhan, H. (1964). Understanding Media: The Extensions of Man. NY, McGraw-Hill.<br />

McMahon, M. and Pospisil, R. (2005). Laptops for a digital lifestyle: Millennial students and wireless mobile<br />

technologies. Proceedings of ASCILITE 2005. (Retrieved November 10, 2010, from:<br />

http://www.ascilite.org.au/conferences/brisbane05/blogs/proceedings/ 49_McMahon%20&%20Pospisil.pdf)<br />

Oblinger, G. (2003). Boomers and gen-Xers, millennials: Understanding the “new students”. EDUCAUSE Review,<br />

38(4), pp.37-47.<br />

Pachler, N., Bachmair, B. and Cook, J. (2010). Mobile Learning: Structures, Agency, Practices. New York: Springer.<br />

Palfrey, J., and Grasser, U., (2008). Born Digital: Understanding the First Generation of Digital Natives. NY, Basic<br />

Books.<br />

Prensky, M. (2001). Digital Natives, Digital Immigrants. On the Horizon, 9(5), pp.1-5.<br />

Rickard, W. and Oblinger, D. (2003). Higher education leaders symposium: The next-generation student. Redmond,<br />

WA. (Retrieved November 10, 2010, from http://www.educause.edu/ir/library/pdf/NLI0425a.pdf).<br />

TecMinho – mLearning –The role of mobile learning in <strong>European</strong> Education. (Retrieved November 10, 2010, from:<br />

http://www.tecminho.uminho.pt/showPage.php?url=el_projs_mLearning.html&static&zid=6).<br />

Thomas, S. (2005). Pervasive, persuasive eLearning: Modeling the pervasive learning space. Proceedings of the 3rd<br />

international conference on pervasive computing and communications workshops (PERCOMW’05). Kauai<br />

Island, Hawai’i: IEEE Computer Society), pp. 332-336<br />

Thomas, S. (2005). Pervasive, persuasive eLearning: Modeling the pervasive learning space. Proceedings of the 3rd<br />

international conference on pervasive computing and communications workshops (PERCOMW’05). Kauai<br />

Island, Hawai’i: IEEE Computer Society, pp.332-336.<br />

Tori, R. (2010). Educação sem distância: As tecnologias interactivas na redução de distâncias em ensino<br />

aprendizagem. Editora Senac - São Paulo, pp.218.<br />

Valkenburg, P., Peter, J. and Schouten, A. (2006). Friend networking sites and their relationship to adolescents’ wellbeing<br />

and social self-esteem, CyberPsychology and Behavior, 9, pp.585-590.<br />

340


A Framework for Information Systems Innovation: A Case of<br />

Competitive Intelligence in Organisations<br />

Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

Tshwane University of Technology, South Africa<br />

phathuts@gmail.com<br />

iyamut@tut.ac.za<br />

Abstract: Knowledge is used as a focal factor for competitive advantage, through effective and efficient<br />

performances by employees in many organisations. As a result, knowledgeable employees are expected to share<br />

their knowledge with others to increase innovation within the organisation. Unfortunately, this is not always the case.<br />

Generally, employees behave differently within an organisation. No organisation has total control of its employees’<br />

behaviour and actions. The behaviour and action, which has impact on how knowledge is shared is influenced by<br />

many factors such as fear of losing jobs, power to negotiate personal interest. These factors influence the<br />

deployment of CI were found and interpreted and conclusion was made from those interpretations. The study aimed<br />

at establishing the impact of Competitive Intelligence (CI) on Information systems (IS) innovation products and<br />

services in organisation. A case study was conducted, using a financial organisation. Innovation-decision process<br />

was applied in the data analysis.<br />

Keywords: competitive intelligence (CI), diffusion of innovation (DoI), Information systems (IS), innovation<br />

1. Introduction<br />

Companies keep competing against each other using products and services, making competitive<br />

intelligence (CI) an important tool in the development of strategy in the organisations. The importance of<br />

competitive intelligence is attributed to its contribution to technological knowledge and intelligence, and it<br />

use for the analysis of information systems innovation in organisations. It should be pointed out early that<br />

innovation and technology are often taken in a similar light as asserted by Rogers (2003: p12) that ‘we<br />

often use the word “innovation” and “technology” as synonyms.’ And as such CI is primarily intended to<br />

be used for the state of art, technological trends and challenges, with a strategic vision on<br />

competitiveness and customers (Ashton & Klavans, 1997; Fleisher & Bensoussan, 2003). Competitive<br />

intelligence therefore can be considered as a tool for innovation process, observation of market, analysis<br />

of strategic behaviors of both competitors and customers, including their values, expectations and needs<br />

(Krücken-pereira et al., 2001).<br />

An innovation according to Rogers (2003: p12) will be considered as ‘an idea, practice, or object that is<br />

perceived as new by an individual or other unit of adoption.’ A unit of adoption in this case could be the<br />

organisation, a society and also a target market.<br />

Competitive Intelligence offers a real strategic advantage for many businesses (Stephen, 2006). Gilad<br />

(2000) argues that some of the largest corporate organizations have a dedicated CI department, while<br />

smaller businesses often practice CI on an ad hoc basis. This they do so by informally collecting<br />

information from a variety of internal and external sources, such as the Internet, trade shows conferences<br />

and networking meetings. Competitive Intelligence is of importance to many businesses mainly because<br />

it helps to formulate strategy, as well as make informed decisions.<br />

The CI is deployed with the intention to better, coordinate internal processes and activities of<br />

organizations, primarily, to reach market more effectively. Gathering people, the logic and the physical<br />

architecture around common purposes provide individuals with the information they need to expand their<br />

own knowledge (Malhotra, 2000; Hoven, 2001). This approach help to build high performance teams in<br />

the organisation. This indeed, is the foundation of the integrated organisation, where the information<br />

technology is capacitating technological innovation.<br />

2. Research methodology<br />

The main focus of the study was to investigate and understand the impact of competitive intelligence on<br />

information systems innovation products and services in organisations. The approaches and methods<br />

employed in the study include the case study, qualitative research method, and semi-structured interview<br />

approach. The data was analysed, using the Innovation-decision process from the perspective of DoI<br />

theory<br />

341


Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

The case study was adopted primarily because it is an approach that assists to achieve a deep<br />

understanding of a specific phenomenon. According to Cooper and Schindler (2006), the case study is an<br />

approach which combines individual and (sometimes) group interviews with record analysis and<br />

observation; used to understand events and their ramifications and processes. Hofstee (2006) argued<br />

that the case study approach is useful when detailed knowledge is required of any particular case. A<br />

pseudonym name, “Divhetsheleni” was used to represent the case, the organisation used in the study.<br />

The participants were codified from DV_LA001 to DV_LA012 to adhere to the ethical consideration as<br />

agreed with the organisation and the university.<br />

The qualitative research method was selected because of the nature of the study. The qualitative study<br />

allows data to be gathered from multiple sources (Yin, 2009). In other words it is not limited to one source<br />

of information.<br />

The research followed a structured interview schedule so that the process does not lose focus. Wellman<br />

and Kruger (2001: 160) assert that “In a structured interview, the interviewer puts a collection of<br />

questions from a previously compiled questionnaire to a respondent face to face”. This was<br />

advantageous for data collection because it made it possible to explain the questions that were not<br />

understood by the respondent and there was chance to further probe responses.<br />

2.1 Diffusion of Innovation (DoI)<br />

The Innovation-decision process from the perspective of DoI theory was employed in the data analysis. In<br />

DoI theory, technological Innovation is communicated through particular channels, over time, among the<br />

members of a social system (Rogers, 2003). The theory is concerned with the manner in which a new<br />

technological idea, artefact or technique, or a new use of an old one, migrates from creation to use.<br />

According to Rogers (2003), the Innovation-decision process involves five steps. As shown in Figure 1<br />

below, the process include: (1) knowledge, (2) persuasion, (3) decision, (4) implementation, and (5)<br />

confirmation. These stages typically follow each other in a time-ordered manner. The stages are briefly<br />

described below.<br />

Figure 1: Diffusion of Innovations (Rogers, 2003)<br />

342


Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

The Innovation decision process characterized as a process that occurs while individuals participate in a<br />

series of actions related to decisions (Rogers, 1995). Knowledge occurs when individuals are aware of<br />

the Innovation and gain understanding of its functions. Persuasion is when individuals or decision-making<br />

units exhibit favourable or unfavourable behaviour toward the Innovation. Decision indicates when the<br />

individual or unit decides to adopt or reject the Innovation. Implementation occurs when the individual or<br />

unit decides to use the Innovation. Confirmation occurs when decision makers confirm or reject their<br />

decision to adopt the Innovation (Rogers, 1995).<br />

3. Data analysis<br />

Using the Innovation-Decision Process (Rogers, 2003), the analysis of the data is presented as follows:<br />

3.1 Knowledge<br />

Knowledge involves management efforts - from identifying needs to delivering intelligence products and<br />

services to consumers and clients. Similarly, innovations are evaluated, monitored and controlled using<br />

knowledge. In the organisation, knowledge was applied in the development of innovation, which was<br />

supported by policy, and generated new requirements. Through knowledge the organisation could<br />

innovate, as well as plan how to utilise the products and services from innovation. According to<br />

DV_LA003 (p4:30-31), Knowledge in the organisation is assessed by questions, and, they are<br />

categorised into needs, such as intelligence problem, decisions, risk metrics (risk tolerance level) and<br />

measurement methods. The organisation was challenged by the knowledge to innovate, primarily, for<br />

competitive advantage.<br />

The organisation focuses on the innovations of security tools and approaches. DV_LA004 (p13:100-101)<br />

explained that there is a growing use of Competitive Intelligence in the financial services sector and<br />

mentioned that the organisation has spent considerable resources to protect their businesses from<br />

infiltration of their systems by other forces.<br />

According to DV_LA003 (p7:45-46), Knowledge is a vital tool, as such, and the organisation conducts<br />

surveys to assess employees' knowledge on the deployment of CI. The organisation worked hard on<br />

communicating with staff, ensuring strategies were clear, addressing performance and remuneration<br />

issues and motivating people. A number of employees, within the organisation, did not have the<br />

knowledge regarding some of their Information Systems innovation Product and Service offerings. This<br />

was attributed to factors such as inadequate skills; no clear consequences of not meeting performance<br />

standards; lack of alignment with strategy and a reward system that fails to motivate properly.<br />

If employees are not informed, or are knowledgeable about products and services it creates risk to both<br />

individuals and the organisation. DV_L003 (p7:62-63) explained that everyone in the organisation are<br />

aware of the strategy and possesses knowledge of the products and services which are deployed in the<br />

organisation.<br />

3.2 Persuasion<br />

At the Persuasion stage, a person’s main type of thinking is affected by, or related to, feelings. Some<br />

employees in the organisation develop an attitude towards certain Innovation and are psychologically<br />

involved with Information Systems Products and Services. According to DV_LA002 (p5:55-56), there is a<br />

marketing team which responsible innovation, and persuades the Board of Directors for approval. Post<br />

approval, the rest of the employees is persuaded on implementation and use of the innovation.<br />

Organisations use incentives and bonuses as tools to persuade employees to perform their tasks, and<br />

align with the organisational innovative strategy. An employee stated that when I am working on any<br />

project, I know that I should do my ultimate best to make sure the project becomes a success. The lack of<br />

some knowledge, on my part, in being able to deliver successfully on the project, affects my incentives<br />

and can, sometimes, also affect everyone involved on the project. DV_LA006 (p16:172-173).<br />

Some employees felt that only using the marketing team is exploitative because of the rest of the team’s<br />

limited perception of the innovation, and reality. Not having a say in the matter may serve as an indicator<br />

of the message’s influence on this stage. However, DV_LA001 (p2:14-15) believes that, by having one<br />

team focusing on finding new ideas to innovate, it eliminates time waste and also cuts cost; inevitably<br />

speeding up the delivery process. DV_LA004 (p8:83-84), commented that there is a perception that<br />

people who work in the field might have first information of latest innovation, but often choose to remain<br />

343


Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

silent because it’s not their place to come up with new ideas. The employee added that it then becomes<br />

more difficult or easier to persuade these people to adopt innovation in the organisation, as they already<br />

know more about it. This depended on the type of findings they have about the innovation, as a result of<br />

their interaction with clients.<br />

Another interviewee, DV_LA005 (p9:148-149) commented that Key Performance Indicators (KPIs) is<br />

good approach to periodically assess the performance of the individuals, business units and the<br />

organisation at large. The employees further explained that in the organisation, KPIs is defined in a way<br />

that is understandable, meaningful, and measurable. They are rarely defined in such a way such that<br />

their fulfilment would be hampered by factors seen as non-controllable by the organisations or individuals<br />

responsible. The KPI ensure that each employee knows the organisation’s expectations, as well as the<br />

output required by the customers.<br />

3.3 Decision<br />

Many of the individuals in an organisation do not adopt innovated Information Systems Products and<br />

Services without proper testing and evaluation. The latter was carried out in relation to the usefulness<br />

and fit in the organisation’s environment. Some individuals used partial trial to gain better understanding<br />

of innovation and deployment of products and services in the organisations.<br />

One of the interviewees, DV_LA008 (p6:17-18) provided the following explanation; some employees in<br />

the organisation are policy decision-makers and initiate requirements for CI products and services. These<br />

decision-makers are the recipients of the end Products and Services of CI through Information Systems<br />

Innovation. The decisions are based on information and knowledge, and sometimes lead to the levying of<br />

more requirements; thus triggering the Intelligence Cycle. After finishing a cycle, a new set, or improved<br />

set, of template will be produced that will be used in monitoring and identifying risk.<br />

According to a Director, DV_LA002 (p5:43-44), It is common to find deployments filled with projects that<br />

have attractive returns on investment but do not move the needle on performance parameters that matter<br />

in the marketplace. In selecting Products and Services to innovate and build a portfolio of improvement<br />

initiatives, clear linkage to strategic priorities needs to be established; not just at the outset, but on an ongoing<br />

basis. When making decisions, one of the most seductive pitfalls is to become comfortable with<br />

embracing innovations that are, in themselves, attractive without paying sufficient attention to overall<br />

optimization.<br />

Although finance could be viewed as the artery of a firm and an important indicator of management<br />

direction, management mindset encompasses more variables than only financial decisions. The<br />

organisation’s brand, strategy, employees and customers play a major role when making a decision to<br />

innovate. The brand has to define the organisation through the innovation. The innovation has to align<br />

with strategy, the employees’ need for knowledge and skills to deploy the CI innovation, and the<br />

customer has to get the best Product and Services that will help solve their problems. DV_LA001 (p3:19-<br />

20).<br />

3.4 Implementation<br />

Each division at Divhetsheleni has its own strategic team that deals with implementing the strategy of the<br />

division, making sure it aligns with the organisational strategy. All CI implementation is defined by the<br />

same vision and goals in order to improve innovation, as well as to create a competitive brand. The<br />

strategy team ensures that Information Systems Innovation aligns with the organisation’s strategy.<br />

However, the organisation had a Marketing Department, which was responsible for innovations that<br />

needed to be implemented in all the various divisions of the organisation. According to one of the<br />

employees, (DV_LA001 p2:18-19), Implementation of a new innovation was about being prepared to<br />

measure the organisation's performance consistently, constantly, to recognise weaknesses and, indeed,<br />

be willing to address those weaknesses. This begins with initial measurement, which serves as an<br />

indication of where the organisation is, so that it can determine where it wants to be. The organisation still<br />

follows old traditional ways of implementation, while Information Systems Innovation challenges the<br />

employees to acquire more knowledge, and understanding, of the products and services they deploy.<br />

The organisation readiness to implement becomes a challenge during the innovative process.<br />

A forward–looking organisation seeks to provide value-added offerings, through Information Systems<br />

Innovation, at every stage of its life in order to improve the culture of the organisation. The organisation’s<br />

344


Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

high-performance culture was characterised by openness and trust, with clear accountability for<br />

execution, and the freedom to take calculated risks; thereby constantly raising individual and<br />

organisational performance. This was expressed by DV_LA005 (p15:204-205).<br />

According to DV_LA002 (p6:64-65), most of the organisation’s successful implementations of a<br />

remuneration strategy, and the alignment of benefits across all staff, are based on job function rather<br />

than grades.<br />

The employee further expressed that, Performance scorecards reflect an integrated approach to<br />

sustainability and are balanced across economic, environmental, social, transformational and cultural<br />

elements. Our commitment to driving high performance is evident in our new organisational structure,<br />

with the separation of support structures from the client-facing clusters to ensure that the business is able<br />

to focus on improving economic performance regardless of economic conditions.<br />

3.5 Confirmation<br />

Confirmation was the recognition, by the organisation, of the benefits of using Information Systems<br />

Innovation; its integration into the organisation’s on-going routine, or promotion of the innovation to<br />

others; as well as the identification of Products and Services to be deployed. This gives the organisation<br />

an opportunity to provide feedback immediately when decisions for tomorrow’s solutions are being made.<br />

According to one of the employees, we establish a team work with our clients from the beginning of the<br />

deployment, DV_LA006 (p18:248-249). This team work between the organisation and the customers<br />

made it possible to test and confirm the proposed innovation.<br />

In addition, another employer, DV_LA001 (p3:29-30) commented that in cases whereby many solutions<br />

required testing, the customer gave the organisation access to people who were experts on the<br />

applications. The collaboration between hardware and software providers, at this early stage, eliminated<br />

the risk normally associated with new technology deployment.<br />

4. Findings<br />

There were some findings from the above analysis of the data. The findings are discussed as follows:<br />

Culture<br />

When an organisation is operating globally, culture always becomes a challenge. The organisation’s<br />

high-performance culture is characterised by openness and trust, with clear accountability for execution,<br />

and the freedom to take calculated risks; thereby constantly raising individual and organisational<br />

performance. For the organisation to deliver on its high performance culture, it must have created an<br />

environment where information is shared openly, whereby its’ people are reward for their skills. In<br />

addition, trust must be created to ensure that employees do not feel used.<br />

Each innovation has performance scorecards that reflect an integrated approach to sustainability. These<br />

are balanced across economic, environmental, social, transformational and cultural elements. In addition,<br />

these innovation performance scorecards assist in building a new organisational structure; separating<br />

support structures from client-facing clusters.<br />

Knowledge Sharing<br />

One of the challenges of an organisation is that of getting people to share their knowledge. In today's<br />

organisation, where so much depends on teamwork and collective knowledge, it is only a handful of<br />

people who have the kind of knowledge with which they can hold their peers (including bosses) to<br />

ransom. Such individuals were Directors or Managers, who were careful and caution of losing trade<br />

secrets. It may be a particular Specialist who has been in the organisation for many years and has built<br />

up his, or her, own unique way of achieving success without perhaps even understanding the deep tacit<br />

knowledge of how they do it. All these poses threat to organisation, when they want to innovate, because<br />

such individuals holds on to the expertise that they have.<br />

Organisation is intended to create a commitment to culture, change, challenge, competition and<br />

cooperation. If, as is often the case, time pressure leads to poor knowledge sharing, then there must be a<br />

345


Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

commitment to allow time for it to happen. Commitment to knowledge sharing must be demonstrated<br />

throughout the organisation.<br />

Repositioning of Brand<br />

The organisation shifts its focus from repositioning its brand to building the brand and the brand promise.<br />

When an organisation innovates, the brand needs to sell the idea. Therefore, the customer has to start<br />

seeing the innovation from brand perspective.<br />

The organisation continues to inspire, motivate and challenge people to make a difference, whilst striving<br />

to become a leader in client-employee understanding and care; across all market segments. The<br />

introduction of a new brand expression, accentuates the intensified, and increased, nationwide above<br />

and below-the-line advertising campaigns to position the organisation’s brand in a more relevant and<br />

approachable manner.<br />

Education and Awareness<br />

Education and Awareness in the organisation encourages and enhances people's participation in<br />

activities aimed at conservation, protection and management of the environment; essential for achieving<br />

sustainable development. The organisation seems to ignore the importance of educating their staff about<br />

Information Systems Product and Services and the path taken by the organisation to deploy Competitive<br />

Intelligence. There appears to be the assumption that seems that all employees are aware of the<br />

deployment process. Employees who were not aware of the innovations which were taking place in the<br />

organisation, could be considered and classified more as expense, than an asset. Their inability to<br />

transfer, or acquire skills, from one other is evident since they are not cognisant of the innovation taking<br />

place within their organisation. Spending the necessary time to educate employees, and making them<br />

aware of CI, will cut organisation costs and reduce time spent doing feasibility studies on the process of<br />

adopting and implementing innovation by an outside partner.<br />

Improve Innovation<br />

The organisation improves innovation across all business clusters in order to create a single view of its<br />

client base. It enables a closer relationship with clients, and it facilitates better co-ordination all actions.<br />

The organisation strives to provide more transactional service offerings, and the driving of primary<br />

investor status, by focusing on further enhancements.<br />

Some divisions, however, were planning on introducing a number of new products, services and business<br />

ventures. A dedicated unit, focused on driving innovative products, has been established. The<br />

organisation will be building on the innovative products that they launch every year; aiming to have<br />

improved its innovation delivery capability significantly.<br />

Inclusiveness<br />

Some of the innovation for process improvement lies with the process owners. This is not a new concept,<br />

but it bears emphasising; as organisations are still not taking this aspect seriously when innovating.<br />

Frequently, process owners are frontline people who do not have the chance to participate in<br />

improvement efforts. While they may not be formally trained in quality tools, their closeness to the<br />

process is a vantage point second to none. The organisation uses their marketing team to come up with<br />

innovation. The same team runs the marketing campaigns and interacts with the customer, forgetting the<br />

very people who actually deploy the process.<br />

Prioritization<br />

One of the most seductive pitfalls is for an organisation to become comfortable with doing Information<br />

Systems innovation projects that are in themselves attractive, without paying sufficient attention to overall<br />

optimisation. It is common to find deployments filled with Information Systems innovation projects that<br />

have attractive returns on investment, but do not move the needle on performance parameters that<br />

matter in the marketplace. In selecting Information Systems innovation projects and building a portfolio of<br />

improvement initiatives, clear linkage to strategic priorities needs to be established, not just at the outset<br />

but on an on-going basis.<br />

346


Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

Creativity is not generally associated with defining the portfolio of Information Systems innovation<br />

projects; however, ongoing re-evaluation can uncover opportunities for innovation. The sum, or where<br />

time and attention is placed, defines an organisation's strategic direction.<br />

The next section presents the interpretation of the findings from the case study.<br />

4.1 Interpretation of the findings<br />

The findings from the analysis are regarded as the factors which influence the Innovation of Competitive<br />

Intelligence (CI) within the Information Systems environment. The Products and Services of CI manifest<br />

into many other components, as depicted in Figure 2 below. The discussion that follows explains each of<br />

the components.<br />

Figure 2: Factors influencing competitive intelligence on information systems innovation<br />

Accessibility – This refers to the organisation’s ability to access information needed for the making of<br />

public decisions as well as its ability to share the economic status of the organisation with its<br />

employees. It also speaks to the sharing of knowledge and expertise by employees, with their peers,<br />

without hiding any information for fear of any threat that might arise.<br />

Power - This is often a cop out by managers, or change agents, who are not adequately addressing<br />

the human factors or motivational aspects. In today's organisation, where so much depends on<br />

teamwork and collective knowledge, it is only a handful of people who have knowledge for which they<br />

can hold their peers (and bosses) to ransom. It might be the owner or manager of a small<br />

organisation not wanting to lose trade secrets; it may be a particular specialist who has been in the<br />

organisation, many years, and built up his or her own unique way of achieving success without<br />

properly understanding the deep tacit knowledge of how they do it.<br />

Enrolment - An individual may have knowledge used in one situation but be unaware that other<br />

people, at other times and places, might face similar situations. Additionally, knowledge derived for<br />

one purpose may be helpful in totally different contexts; or it may be a triggered for innovation. Many<br />

innovative developments come from making knowledge connections across different disciplines and<br />

organisational boundaries.<br />

Management - There is pressure on productivity, on deadlines, and it's a general rule that the more<br />

knowledgeable you are, the more there are people waiting to collar you for the next task.<br />

Management ensures that lessons learnt by individuals and groups are captured into the knowledge<br />

database for sharing. .<br />

Competitive Advantage – This is the strategic advantage one organisation entity has over its rival<br />

entities within its competitive industry. Achieving Competitive Advantage strengthens and positions<br />

an organisation better within the business environment.<br />

In order for an organisation to develop a competitive advantage it must have resources and capabilities<br />

that are superior to those of its competitors. Without this superiority the competitors can simply replicate<br />

what the organisation was doing, and any advantage disappears quickly.<br />

Framework for Information Systems Innovation<br />

From the analysis and findings, the case study provides insight to the different characteristics of<br />

Competitive Intelligence (CI) Products and Services Innovations within the Information Systems<br />

347


Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

environment. This is illustrated in Figure 3. To have a full understanding, the discussion that follows<br />

should be read with the figure 3.<br />

Figure 3: Competitive intelligence products and services innovations within the information systems<br />

environment<br />

Information Systems innovation begins with the organisational environment, whereby there is top-down<br />

support for project management practices and investment, by the organisation, in the innovation process.<br />

In rolling out Products and Services, implementation by the organisation focuses on doing so correctly,<br />

rather than simply doing what it takes to complete the CI deployment as quickly as possible, at a<br />

minimum cost.<br />

Organisations that are strong in Adaptability and Involvement have an edge in innovation and creativity,<br />

while organisations excelling in Mission and Consistency have a high measure of stability, return on<br />

investment and return on sales. Organisations measuring high in all components have a dramatic<br />

financial advantage over organisations that are weak in these areas. Organisations at the bottom perform<br />

just as one would expect: They are sluggish, wasteful and out of touch with their customers.<br />

There is reliance, by the organisation, on the Key Performance Indicators (KPI’s) to identify or measure<br />

the acceptance of innovation by a client. KPI, for an organisation, is a tool that persuades decision<br />

makers as a result of the reaction of the customer towards the organisation’s innovation. It also measures<br />

an employee’s performance on certain tasks related to the innovation.<br />

If successful implementation of Products and Services contributes directly to strategic objectives and the<br />

bottom line, Information Systems Innovation should lead to organisational success. Innovation requires<br />

an investment in resources, therefore responsibility lies with senior management to ensure that<br />

investments are made wisely and add value to the organisation as a whole. Unsuccessful CI<br />

deployments translate into poor investment and, ultimately, senior management is accountable for<br />

projects not reaching the objectives of the organisation.<br />

In conclusion, the brand of an organisation brand is a key intangible attribute by which they compete. Its<br />

main objective is to support differentiators; and reposition the company as being significantly different<br />

from any other financial organisation. In so doing, potential clients are drawn to the brand which,<br />

ultimately, encourages them to choose their organisation above that of their competitors. The role of<br />

competitive intelligence on information systems innovation is widely revealed. It was better understood<br />

that while competitive intelligence is overemphasized as revolutionary customers focused, information<br />

systems products and services still remain challenging. It was also understood that not all organisations<br />

that deploy competitive intelligence are responsible for the innovation. Lack of knowledge sharing and the<br />

organisational culture were found to be important factors for the deployment of competitive intelligence<br />

products and services in the organisations.<br />

From the findings and interpretation of the case study on the impact of Competitive Intelligence on<br />

Information Systems products and services innovations in organisations, a Framework was developed,<br />

Figure 4. The Framework is aimed at providing direction for organisations during Information Systems<br />

348


Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

innovation, the stages on the Framework encompasses the activities involved in Information Systems<br />

innovation.<br />

Figure 4: Framework for information systems innovation<br />

The Information Systems it’s composed of five different stages that organisations need to follow for<br />

deployment of CI products and services. Each Stage has columns that are categories by different<br />

aspects of innovation on a row level. All innovation stages or phases actives, roles and responsibilities,<br />

requirements and information systems inputs for the deployment of a successful competitive intelligence<br />

products and services has been illustrated in this framework.<br />

5. Conclusion<br />

Knowledge sharing is critical to Information System Innovation in the organisation that deploys it, making<br />

it very useful for competitive advantage. The study helps managers to gain better understanding of how<br />

knowledge sharing influences products and services in the organisations.<br />

Sometimes the effect of Competitive Intelligence on the innovative process is not obvious, but it does<br />

exist because the companies compete, and are challenged with customers’ needs. Therefore it is<br />

strategically important to equip as many as possible employees with enough knowledge to carry out the<br />

customers’ demands. Regardless of how knowledge is acquired, Competitive Intelligence deployment<br />

relies on the knowledge individuals and group have about the information Systems products and<br />

services.<br />

References<br />

Ashton, W. B. Klavans, R. A. (1997). Keeping abreast of science and technology: technical intelligence for business.<br />

United States: Battelle Press.<br />

Cooper, D. & Schindler, P. (2006). Business Research Methods, McGraw-Hill, New York.<br />

Den Hoven, J. Van (2001). Information Resource Management: Foundation for knowledge management.<br />

http://www.brint.com/members/01060524/ irmkm/irmkm_1.html.<br />

Fleisher, C.S. (2003). "Competitive Intelligence Education: Competencies, Sources and Trends," Information<br />

Management Journal, March/April, 56-62.<br />

Gilad, B. (2000). "An Ad Hoc, Entrepreneurial CI Model," Competitive Intelligence Magazine, vol. 3, no. 4, pp. 33-39.<br />

Gregor, S. & Jones, K. (l999). Enabling Electronic Commerce in Agribusiness: The Initial Stages of a Diffusion<br />

Process (where - Collecter or Bled?).<br />

Grover, V. (1997). An extension of the Tri-core model of information systems innovation: Strategic and technical<br />

moderators. <strong>European</strong> Journal of Information Systems, 6(4), 232-242.<br />

349


Phathutshedzo Nemutanzhela and Tiko Iyamu<br />

Haag Stephen, (2006). Management Information Systems for the Information Age. Third Edition. McGraw-Hill<br />

Ryerson.<br />

Hofstee, E. (2006). Constructing a Good Dissertation: A Practical Guide to Finishing a Master's, MBA or PhD on<br />

Schedule, EPE Publishers, South Africa.<br />

Hovav, A., Patnayakuni, R. & Schuff, D. (2001). Internet Technology Diffusion: Adoption ofIPV6 Global Cooperation<br />

in the New Millennium, 9th ECIS, Bled, Slovenia.<br />

Krücken-Pereira, L.; Debiasi, F.; Abreu, A. F., 3, jul (2001).Technological innovation and competitive intelligence: an<br />

interactive process. READ - Electronic Journal of Directors, v.7, Available at:<br />

. Accessed 19 September 2003.<br />

Malhotra, Y. (2000). Knowledge Management & New Organization Forms: A framework for Business Model<br />

Innovation. Information Resources Management Journal. Jan-Mar., vol.13, no.1, pp.5 – 14.<br />

Rogers, E. M., (1995). Diffusion of Innovations. 4 th ed. New York: Free Press.<br />

Rogers, E.M., (2003). Diffusion of innovations (5th ed.). New York: Free Press.<br />

Welman J. C. & Kruger S.J, (2001). Research Methodology: for the Business and Administrative sciences, 2 nd Ed.<br />

Oxford University Press, South Africa.<br />

350


RFID: A Review of its Relevance and Application in South<br />

African Retailing Systems<br />

Jonathan Oni, Edward Dakora and Vesper Owei<br />

Cape Peninsula University of Technology, Cape Town, South Africa<br />

jo45ng@yahoo.com<br />

naadakora@gmail.com<br />

OweiV@cput.ac.za<br />

Abstract: The advent of Radio Frequency Identification (RFID) has enabled identifying an object using a radio<br />

signal. RFID has become the talk of the day when world retail giants including Wal-Mart, Tesco, Target, and<br />

Albertsons announced their implementation of the technology to ensure efficient supply chain management, and,<br />

naturally, demanded that their suppliers meet RFID requirements. Today, these retailers boast some of the best<br />

supply chain systems and management in the industry worldwide. The increase in adoption and application of RFID,<br />

therefore, seems inevitable. It has been envisaged that the near future will see almost all new products sold to<br />

consumers having an RFID tag that can allow them to be remotely and uniquely identified and tracked. The potential<br />

impact of RFID on the South African retail industry is significant. This technology is mostly employed in supply chain<br />

management and customer relationship management. In supply chain management, RFID can benefit retailers in<br />

many ways apart from the mere identification of products and suppliers, such as monitoring the production process<br />

and the sequence of activities from the material production to delivery of the final products. Preliminary studies reveal<br />

that while South African retailers are aware of RFID and are interested in its adoption, few are actively engaged in<br />

pilot tests or planned implementations. Although, they are aware of its benefits, the relevance and its application in<br />

the South African retail sector is still limited. This paper reviews the relevance of RFID and the application in South<br />

African retail system.<br />

Keywords: RFID, RFID relevance, RFID application, South Africa, retail system<br />

1. Introduction<br />

RFID have led to benefits such as cost reduction, low shrinkage, and increase in value and quality of<br />

products. Fitzek (2003) noted that the benefits of RFID may manifest throughout the supply chain, the<br />

primary benefits are evident at the retail end, including marketing and customer service provision.<br />

Although, the South African retail system is experiencing a slow pace in adoption and application, the<br />

benefits of RFID is enormous.<br />

This paper reviews the relevance of RFID and the application in South African retail system. The paper<br />

further explores in greater depth the factors that influence the application of RFID in the South African<br />

retail system. The findings and their implications will be discussed, and the limitations and ideas for future<br />

research will be outlined. The paper concludes with a discussion on the application of RFID and its<br />

relevance to the South Africa retail system. The rest of the paper is structured as follows: The next<br />

section is a literature review of RFID global perspective and adoption in South Africa. This is followed by<br />

the research method in section 3. Section 4 discusses the analysis and findings of the research method.<br />

The discussions and limitations of the research are presented in section 5. The paper concludes in<br />

section 6 with a summary of the paper and some positive thoughts on the way forward for a successful<br />

application of RFID in the retailing sector of South Africa.<br />

2. Literature review<br />

2.1 ICT in emerging economies<br />

The ICT economy report hailed this technology as what fuelled the “strong wave of innovation that<br />

transformed the global economy during the last quarter of the 20 th century” (UNCTAD, 2007). The<br />

Internet and World Wide Web as part of ICTs are being used by businesses of all kind around the world<br />

in the management of their day-to-day operations and transactions, and, therefore, forms an integral part<br />

of their marketing strategies (Sagi, Carayannis, Dasgupta & Thomas, 2004). The growth of the ICT sector<br />

has transformed many economies around the world, making the drivers of growth to be more information<br />

and knowledge based rather than the traditional natural resource dependency (ITU, 2007). However, the<br />

said report also acknowledges the disadvantages of developing countries including limited infrastructure,<br />

human capital, policy incentives, investments and high cost of services which hamper their growth in<br />

ICTs.<br />

351


Jonathan Oni et al.<br />

Yet, it has been argued that the arrival of ICT revolution in Africa represents one aspect of globalisation<br />

on the continent (Carmody, 2009). Among ICTs, mobile phones have been the most accepted and<br />

adopted on the continent, with growth and penetration rates reaching alarming proportions, although this<br />

rate may vary from country to country. The high penetration rate of mobile technology and mobile phones<br />

in African countries represent a significant contribution to economic development, and poverty reduction.<br />

The global IT report (Geiger & Mia, 2009) affirms the positive impact of Mobile phones penetration on<br />

economic growth, and its potential to reduce poverty and bridge the digital divide. Moreover, the<br />

completion of the SEACOM’s fibre optic cable system, connecting Africa to the rest of the world and<br />

providing faster internet connectivity is a major boost to the development of ICTs on the continent.<br />

However, there is still a limitation on individual ownership of ICTs equipment and Internet connectivity in<br />

developing economies, and, therefore, ICTs supported business model that are proven to be successful<br />

the developed world may not be readily application in the developing world situation (Roztocki &<br />

Weistroffer, 2009). ICT based innovations that are relevant to the developing market situation might be<br />

necessary. For example, in the case of RFID technology and its application retailing, the passive RFID as<br />

will be discussed later, might be more user friendly to retailers in developing country markets like South<br />

Africa. In South Africa, retailers are seeking more efficiency in the mist of the economic downturn by<br />

implementing new ICTs including Web-based portals, messaging platforms and wireless technologies in<br />

their operations (Dos Santos & Baccialoni, 2009).<br />

2.2 Global adoption of RFID<br />

With the hype of RFID, Chen (2004); Riggins and Slaughter (2006) suggest that the market is predicted<br />

to grow by round about 20% each year and will reach 2-3 billion US dollar in 2008.Though, after many<br />

years of hyping the technology, the actual adoption lags behind the optimistic expectations. According to<br />

Matta and Moberg (2006), besides all attention which has been paid to the RFID technology, the pace of<br />

actual or planned RFID adoptions by companies in their supply chains remain low to moderate. Various<br />

and recent studies have reported adoption rates only between 7-15% (Brown & Bakhru, 2007);<br />

(Leimeister & Knebel, 2007); (Matta & Moberg, 2006).<br />

It has been envisaged that the near future will see almost new products sold to consumers having an<br />

RFID tag that can allow them to be remotely and uniquely identified and tracked (Cazier, Jensen & Dave,<br />

2008). RFID tags are said in Cazier’s et al. work to bear products information including but not limited to<br />

product type, price, place of origin, and place of purchase. This information can be read using any<br />

appropriate scanner. Two types of RFID tags exist according to Angeles, (2005); Cazier et al.:<br />

Active tags rely on a power source (battery) to function, and this power source makes it possible for<br />

the tags to broadcasts “their own signal over varying distances, depending on the potency of their<br />

power source and range of frequency.” Nevertheless, active tags are said to be very expensive and<br />

are therefore, not popular in business environments; their applications are more prominent in military<br />

and other such environments.<br />

Passive tags use no battery, “sit idle until passed near a reader that emits radio waves” (Haag, 2007)<br />

and are economical to produce. They are not capable of broadcasting their signal due to their lack of<br />

power source; they are common with consumer goods.<br />

Figure 1: A typical RFID system (Robert, 2006)<br />

Since the biggest benefits of RFID can be seen in the supply chain sector (Curtin et al., 2006), it is logical<br />

to study its adoption from an organizational perspective as opposed to individual adoption. Applying<br />

various database, Schmitt and Michahelles (2009) presented a result of studies conducted on the<br />

adoption of RFID. The reviews were adapted and summarized in Figure 2. The first column indicates the<br />

352


Jonathan Oni et al.<br />

author(s) followed by the research topic. The third column shows the used research methodology while<br />

the fourth column summarizes the results of the study especially the identified adoption factors.<br />

Figure 2: RFID adoption studies adapted from Schmitt and Michahelles (2009)<br />

2.3 The South Africa retailing system<br />

South Africa appears to have the oldest formal retailing system in Africa, with an early experience of the<br />

supermarket revolution in the developing world, the first wave in the 1990s. This system has advanced<br />

353


Jonathan Oni et al.<br />

over the years, with a lot of takeovers, mergers and acquisitions, conglomerations and expansions. To<br />

date, the country has retail companies grown into chains and groups, and become regional giants. For<br />

example, Shoprite has grown to become the largest food retailer in the continent over the past decade,<br />

due to its Africa-wide expansion drive.<br />

Moreover, the South African retail system also meets international standards, and, in a few instances,<br />

South African retailers have won internal awards for innovation. Woolworths (SA) has won twice in a roll,<br />

the coveted award as the worlds “Responsible Retailer of the Year” at the World Retail Congress 2010;<br />

beating Cargills Food City, Kingfisher, New Look, 0 Boncario and Teknosa to it this year (The Word Retail<br />

Congress, 2010). This is due to Woolworths “Good Business Journey” initiative, for sustainable farming<br />

and organic food production, locally. At the same event, Truworths International (SA) won the Emerging<br />

Market Retailer of the Year. All these retailers have employed numerous technologies and innovative<br />

strategies from, customer relationship management to supply chain management. Most of the clothing<br />

retailers like Truworths and the Foschini have credit account facilities allowing customers to shop now<br />

and pay later, in some cases it take just an SMS to open such an account. There are substantial use of<br />

scanning systems and devices at the stores and distribution centres some of which would be RFID.<br />

However, it appears the use adoption and use of RFID scanning systems for tracking products through<br />

the supply chain to stores is limited in the South African retail system. The annual reports of the major<br />

retailers make mention of the use of sophisticated computers systems and supply chain management<br />

systems, especially in the case of the supermarket chains, but they are silent on the specific use of RFID.<br />

2.4 Relevance of RFID to South African retailing<br />

The potential impact of RFID on the retail industry is significant. This technology is mostly employed in<br />

supply chain management and customer relationship management. In supply chain management, RFID<br />

can benefit retailers in many ways apart from the mere identification products and suppliers, such as<br />

monitoring the production process and the sequence of activities from the material production to delivery<br />

of the final products (Chang, Hung, Yen & Chen, 2008). Chang et al. also point out that retailers in<br />

developed countries such as America and Japan have employed some unique characteristics of RFID to<br />

enhance their global supply chain management and efficiency. This also leads to other benefits such as<br />

cost reduction, and increase in value and quality of products, Chang et al. indicates.<br />

With South Africa’s already sophisticated retail system, and their drive to expand into the rest of the<br />

African continent, the management and movement of products to stores in different countries have a<br />

critical issue. RFID appear to be the technology needed to help trace, manage the handling of products<br />

all the way from the manufacturer to the point of sale in retail store. Despite the potential usefulness of<br />

RFID, the adoption and application of this technology among South African retailers seem very limited.<br />

Brown and Russell (2007) in their study of RFID adoption in South African retailers found no traces of its<br />

adoption; the most progressive retailer had only began talking about it for possible future exploration. To<br />

date, the Annual reports of the major retailers in South Africa are still silent on the adoption and<br />

application of RFID.<br />

Moreover, the benefits of RFIDs are not limited to supply chain, but are enjoyed by other parts of<br />

retailing, including marketing and customer service provision. In this line of thinking, Eckfeldt (2005) and<br />

Uhrich et al., (2008) identified some three ways in which RFID can provide value to consumers including<br />

convenience, improved service, and increased customer satisfaction, which together can result in<br />

competitive advantage and stimulate demand.<br />

Figure 3 illustrates a typical produce supply chain, and with the application RFID, the retail at the store at<br />

can track the goods right from the point of production through to the final consumer, even after purchase.<br />

3. Research methodology<br />

The character of this paper calls for the participation of different stakeholders within the South African<br />

retail sector. This starting point directs the research primarily towards the use of a qualitative research<br />

method to be applied. An accurate description of qualitative research is that it is concerned with analysing<br />

human action in terms of meanings (Ezzy, 2002). Qualitative research is noted by the significance of<br />

producing detailed or thick descriptions of the social settings being investigated. Neuman (2000) agrees<br />

with Myers (1997) that qualitative data are in the form of text, written words, phrases, or symbols<br />

describing or representing people, actions, and events in social life. In line also with the various<br />

354


Jonathan Oni et al.<br />

description of qualitative research, Blaikie (2000 p.251) assert that qualitative researchers “give a great<br />

deal of attention to describing what might appear to be minute and trivial details of social activity. It is<br />

argued that this is necessary to provide a backdrop within which actions and interactions can be<br />

understood”.<br />

Figure 3: Produce supply chain (source: GS1 (2009:8))<br />

In this paper, the semi-structured interview approach was employed. This involves commencing<br />

interaction as an unstructured interview by presenting the primary overall question, and then using certain<br />

questions to elicit information. The interview questions are designed to provide the in-depth analysis<br />

required and also to allow the authors to probe for further information. One advantage of interviews as<br />

data collection techniques is that they allow the researcher the opportunity to use individual quotes and<br />

refer to specific situation as implied by the interviewee.<br />

3.1 Document analysis<br />

Extensive academic literatures on RFID adoption and implementation were reviewed (see figure 1.2);<br />

including journals on previous related research, as well as white papers. The articles were reviewed and<br />

analyzed and the gaps identified. The aim being to ascertain to what extent RFID is being applied to the<br />

South African retailing sector and its relevance.<br />

3.2 Interviews<br />

The interviewed managers were asked about the relevance of RFID and their application across different<br />

chain stores. They were also asked about their experiences with RFID application in the retail sector of<br />

South Africa. Specifically, they were asked factors that influence the application across various retail<br />

outlets as well as challenges being faced.<br />

The study targeted some retail industry, namely the fast moving consumer goods (FMCG) retailers in<br />

South Africa. The first series of interviews was conducted with Four (4) retail and store managers. Three<br />

(3) additional IT Consultants’ with several years of network technologies and application experience were<br />

interviewed. The interviews were conducted between September and November 2010. Although, two of<br />

the consultants could not be reached for a face-to-face interview due to time constraints, they were<br />

however asked to complete the interview questions at their private time and return their response via<br />

email. To ensure participants of their anonymity, each participant was provided with an informed consent<br />

form outlining the aims of the paper and stating that they will remain anonymous at all times.<br />

The authors were aware of shortcomings of the interview technique such as interview bias, interviewer<br />

effect, and the potential impact of the interviewer characteristics. The interview commenced only after<br />

355


Jonathan Oni et al.<br />

they agreed to participate. Information gained in the interviews were recorded, transcribed and analysed<br />

afterwards.<br />

4. Analysis of findings<br />

Our findings identified several factors that influence the application of RFID across South African retail<br />

sector. The relevance RFID was also outline by the various respondents in the interview. The response of<br />

the interviews are presented and discussed below.<br />

4.1 Cost<br />

Although there are compelling reasons to adopt RFID to replace barcodes as noted by several<br />

respondents, issues of cost have been identified. The ever changing technological know-how brings new<br />

challenges. The software cost, integration and implementations, and supporting infrastructure costs are<br />

very high. The additional cost of ongoing training of staff coupled with the risk involved in loosing these<br />

staff makes the adoption and application less attractive. With the slow pace in buying power of<br />

customers, investing on RFID is considered ‘suicidal’ as noted by one of the respondent.<br />

Cost concerns are consistent with various literatures (Brown & Russel, 2007); (Huber et al., 2007).<br />

Thus, this supports the respondents’ view that costs associated with RFID negatively influence its<br />

adoption and application.<br />

4.2 Technology<br />

The pace of technology adoption in Africa is considered limited due to the cost involved. Although,<br />

innovative organizations encourage adoption of new technologies, the fear of perceived benefits<br />

discourages some, whether they are from within or outside of the organization. In line with this, Asif and<br />

Mandviwalla (2005), states that technology can add to value by reducing costs or enhancing quality. The<br />

technological environment of most retailing sector in South Africa does not encourage the adoption of<br />

RFID. One respondent noted the following, “We do not have the basic tools that will encourage the<br />

application of RFID, even if we are able to afford it, we still do not have the man-power to implement it”.<br />

Other respondent also feel that RFID technology can not be compatible will their organizational<br />

structures. The uncertainties that surround future Government policies regarding storage and use of<br />

information among retailers also hinder the application of RFID.<br />

4.3 RFID readiness - environment<br />

Most of the responders draw attention to the environmental conditions which have impacted on adoption.<br />

The readiness to adopt and implement RFID will greatly depend on the business environment. Real<br />

advantage of RFID can only be gained in the supply chain when suppliers and customers collaborate.<br />

This implies that willingness of both parties to improve their environment and business processes might<br />

facilitate the adoption of a technology that will bring benefits to both of them (Fawcett et al., 2006).<br />

One respondent cited “external forces, internal resources, technical and behavioural issues” as some of<br />

the potential factors influencing RFID application in the retail sector. It can thus be implied that the<br />

relevance of RFID is seen among the different retails outlets but the enabling environment has imparted<br />

negative on its full application.<br />

4.4 IT skills<br />

In South Africa, skills shortage across a range of market sectors especially in information technology and<br />

engineering is a pressing issue (Anderson, 2008; Mulder, 2007). Hence, the non-availability of necessary<br />

IT skills to implement, integrate and maintain RFID greatly impacts on his adoption and application. One<br />

respondent noted “…we can only implement this technology (RFID) if we can boost of the necessary IT<br />

skills, which at the moment, we can no”. Another respondent said “Our head office will already short staff<br />

in the IT department; embracing RFID will compound our already existing problem of lack of skills and<br />

staff”. As reported by Dainty et al. (2005), effort to create sustainable future supply of indigenous skills<br />

must begin with a robust campaign to promote the industry, its occupations and its careers. This process<br />

will help in building and maintaining a talent pipeline through job experience that is not only confined to<br />

home market but also to the global market (Hall & Sandelands, 2009).<br />

356


4.5 Improved delivery and reliability<br />

Jonathan Oni et al.<br />

The respondents’ acknowledge that RFID has enabled substitution of manual processes into an<br />

electronic means. This in turn has resulted in enhanced delivery of goods and services with accuracy and<br />

faster processing capability. Another area of its relevance is the complete acquisition of outgoing<br />

shipments both within the organisation and outside correspondence. One respondent noted that “this<br />

technology has enabled them to trace missing materials from warehouse to the retail outlets”. Perhaps,<br />

this improved delivery and reliability would be apparent with Wal-Mart’s bid to take over Massmart in<br />

South Africa; presenting some retailers an opportunity for RFID to be introduce in a significant way in the<br />

South African retailing system, if, of course, the takeover goes through successfully.<br />

4.6 Increased efficiency in the supply chain<br />

The goal of cost efficiency and fast handling of stocks with the supply chain has been enhanced with<br />

RFID. As large volumes of goods are handled, savings in time in handling these goods become an<br />

important advantage. For example, it has been noted that expediting the loading and unloading of trucks,<br />

reduces the risks of contamination of goods (Shulman, 2001). One respondent agrees that “increased<br />

efficiency in the supply chain using RFID has also enabled them to track products not purchased before<br />

their sell-by date”. This agrees with reports stating that in the <strong>European</strong> grocery sector, products that are<br />

not purchased before their sell-by date are estimated to cause yearly costs of billions of dollars (Leicester<br />

University & Cranfield University, 2001).<br />

4.7 Improved service<br />

Some South African retails stores are already beginning to weigh how RFID-based systems might<br />

improve the overall customer-service and consumer shopping and -spending experience. This for<br />

example can be seen in Prada’s experimenting with high-tech futuristic stores that provide real-time<br />

inventory to sales associates, on-demand multimedia displays of sexy fashions, and dressing rooms<br />

showing customers runway images of models wearing the clothes they’re trying on (Eckfeldt, 2005).<br />

Airlines in South Africa like their counter parts, including Delta Air Lines, are beginning to use RFID<br />

tracking systems to ensure that baggage arrives on time at the correct, specified destinations. These<br />

show that there is improved service satisfaction on both the organisation implementing RFID and its<br />

customers.<br />

5. Conclusion<br />

Although there is substantial interest in the adoption and application of RFID technology by major<br />

retailers around the world, this interest appear comparatively limited in the South African retail system<br />

due to several constraints as discussed earlier. But the situation is not unique to South Africa as most<br />

developing economies have general ICT related issues.<br />

This study has shown that cost, technology, RFID environment readiness, IT skills are factors influencing<br />

the full application of RFID in South African retailing sector. It has also identified that RFID is being<br />

utilized in various sectors to improve delivery and reliability, increase efficiency in the supply chain, and<br />

improve service among organisations and customers especially in the case of global retail chains like<br />

Wal-Mart. As RFID technology progresses in its development, and application and becomes readily<br />

available in the near future, there will be vast opportunities for different organisations to not only embrace<br />

the technology, but also to examine its implications and benefits in a larger scale. Hence, this study might<br />

be useful in promoting the relevance and application of RFID not only in the retail sector of South Africa<br />

but also in other sectors. The information gathered will assist organisations that want to implement this<br />

technology.<br />

To successfully adopt and implement RFID in the retail sector of South Africa, enterprises will have to<br />

resolve issues of cost, technology development, IT skills and environment readiness. As these issues are<br />

addressed, it is certain that RFID will continue to innovate and diffuse, and it will quickly be assimilated<br />

into our daily lives. Though, limited South African firms are employing RFID to improve the efficiency of<br />

their operations and to gain a competitive advantage, it is believed that the near future will see an<br />

increase in this.<br />

6. Limitations<br />

The issues covered in this research are a reflection of the different interviews conducted with various<br />

retail and store managers as well as IT consultants. Consequently, some of these issues may differ from<br />

357


Jonathan Oni et al.<br />

one organisation to another and may not directly reflect a particular RFID adoption and application<br />

approach. A wider and more quantitative study might be necessary in the future to allow for<br />

generalisation.<br />

7. Contribution<br />

The experience gathered from this study will be applied to programs focused on RFID adoption and<br />

application as well as relevance in the South African context. It will assist decision makers in identifying<br />

potential factors that could influence RFID adoption. It is hope that by sharing this research study, it can<br />

stimulate the use of RFID as a tool to improve and gain competitive advantage in the retail sector. This<br />

study has not included all the factors that could influence RFID application or its relevance to the South<br />

African context such as organisation perception, competitive pressure, industry and trade, etc. Thus, this<br />

can be an area for future research.<br />

References<br />

Anderson, B. (2008). "Dire skills gap", Finweek, No.17 April, pp.81.<br />

Angeles, R. (2005). RFID Technologies: spply-chain applications and implementation issues. Information systems<br />

management, Winter: 51-63.<br />

Asif, Z. and Mandviwalla M. (2005). Integrating the Supply Chain with RFID: A technical and business analysis.<br />

Communications of the Association for Information Systems, 15, 393 - 427.<br />

Blaikie, N. W. H. (2000). Designing social research: the logic of anticipation, Polity Press, Malden, MA.<br />

Brown, A., Bakhru, A. (2007). Information Systems Innovation Research and the Case of RFID. In: McMaster, T.,<br />

Wastell, D., Ferneley, E., DeGross, J. (Hrsg.). Organizational Dynamics of Technology-based Innovation:<br />

Diversifying the Research Agenda, Volume 235, 363-376, Boston, MA: Springer.<br />

Brown, I. & Russell, J. (2007). Radio frequency identification technology: An exploratory study on adoption in the<br />

South African retail sector. International Journal of Information Management, 27: 250-265.<br />

Carmody, P. (2009). A new socio-economy in Africa? Thintegration and the mobile phone revolution. IIIS discussion<br />

paper no. 279. URL.<br />

Cazier, J.A., Jensen, A.S. & Dave, D.S. (2008).The impact of consumer perceptions of information privacy and<br />

security risks on the adoption of residual RFID technologies. Communications of the Association of Information<br />

Systems, 23(14): 235-256, September.<br />

Chang, S.I., Hung, S.Y., Yen, D.C. & Chen, Y.J. (2008). The determinants of RFID adoption in the logistics industry -<br />

a supply chain management perspective. Communications of the Assocation for Information System,<br />

23(12):197-218.<br />

Chen, Y.-H. (2004). Getting ready for RFID. OR/MS Today, 31 (3), 30-35.<br />

Curtin, J., Kauffman, R. J., & Riggins, F. J. (2006). Making The ‘Most’ Out Of RFID Technology: A Research Agenda:<br />

For The Study Of The Adoption, Usage And Impact Of RFID. Paper presented at the Information Technology<br />

and Management.<br />

Dainty, A.R.J., Ison, S.G., Root, D.S. (2005). "Averting the construction skills crisis: a regional approach", Local<br />

Economy, Vol. 20 No.1, pp.79-89.<br />

Dos Santos, L.A. & Bacchialoni, A. (2009). ICT in the South African retail and wholesale vertical. White<br />

paper.http://www.bmi-t.co.za/?q=content/sa-ict-retail-and-wholesale-vertical-sector [20 September 2009].<br />

Eckfeldt, B. (2005). What does RFID do for the consumer? Communications of the ACM, 48(9): 77-79, September.<br />

Ezzy, D. (2002). Qualitative analysis: Practice and innovation, Allen & Unwin, Crows Nest, N.S.W.<br />

Fawcett S.E., Ogden J.A., Magnan G.M. and Cooper M.B. (2006). Organizational commitment and governance for<br />

supply chain success. The International Journal of Physical Distribution & Logistics Management, 36, 22 - 35.<br />

Fitzek, D. (2003). Application of RFID in the grocery supply chain: Universal solution for logistics problems in the<br />

CPG industry or a mere hype? Technical Report, Universitat St. Gallen.<br />

Geiger, T. & Mia, I. (2009). Mobile telephony: a critical enabler of networked Readiness?.In Dutta, S. & Mia, I. (eds).<br />

The global information technology report 2008-2009: mobility in a networked world.<br />

http://www.weforum.org/pdf/gitr/2009/gitr09fullreport.<br />

GS1. (2009). Traceability for fresh fruits and vegetables implementation Guide.<br />

http://www.gs1.org/docs/gsmp/traceability/Global_Traceability_Implementation_Fresh_%20Fruit_Veg_i1.pdf [22<br />

November 2010].<br />

Haag, S., Baltzan, P. & Phillips, A. (2007). Management information systems for the information age, Boston, MA:<br />

McGraw-Hill/Irwin.<br />

Hall, J., Sandelands, E. (2009). Addressing South Africa's engineering skills gaps", Education + Training, Vol. 51 Iss:<br />

3, pp.215 – 219<br />

Huber, N., Michael, K., McCathie, L. (2007). Barriers to RFID Adoption in the Supply Chain. IEEE RFID Eurasia,<br />

Istanbul, 1-6.<br />

ITU: International Telecommunication Union. (2007). WORLD INFORMATION SOCIETY REPORT. www.itu.int/wisr.<br />

[04 January 2011].<br />

Jones, P., Clarke-Hill, C., Shears, P., Comfort, D., Hillier, D. (2004). Radio Frequency Identification in the UK:<br />

Opportunities and Challenges. International Journal of Retail & Distribution Management, 32 (3), 164-171.<br />

Lai, K.-H., Ngai, E.W.T., Cheng, T.C.E. (2005). Information Technology Adoption in Hong Kong's Logistics Industry.<br />

Transportation Journal, 44 (4), 1-9.<br />

358


Jonathan Oni et al.<br />

Leicester University and Cranfield University. (2001), Shrinkage – Introducing a Collaborative Approach to Reducing<br />

Stock Loss in the Supply Chain, ECR Europe, research report.<br />

Leimeister, J.M., Knebel, U., Krcmar, H. (2007). RFID as Enabler for the boundless real-time Organisation: Empirical<br />

Insights from Germany. International Journal of Networking and Virtual Organisations, 4 (1), 45-64.<br />

Matta, V., Moberg, C. (2006). The Development of a Research Agenda for RFID Adoption and Effectiveness in<br />

Supply Chains. Issues in Information Systems, 7 (2), 246-251.<br />

Mulder, C. (2007). "Nurture the minority who show potential and desire", Finweek, No.8 February, pp.60.<br />

Myers, M. (1997). 'Qualitative Research in Information Systems', MIS Quarterly, vol. 21, no. 2, pp. 241-242.<br />

Neuman, W. L. (2000). Social research methods: qualitative and quantitative approaches, 4th edn, Allyn and Bacon,<br />

Boston.<br />

Roberts, C.M. (2006). Radio frequency identification (RFID). International Journal of Compute r s and S e c u r i t y, 2<br />

5 (2006) 18 – 26.<br />

Riggins, F.J., Slaughter, K.T. (2006). The Role of Collective Mental Models in IOS Adoption: Opening the Black Box<br />

of Rationality in RFID Deployment. Proceedings of the 39th Hawaii International <strong>Conference</strong> on System<br />

Sciences (HICSS), Hawaii, HI.<br />

Sagi, J., Carayannis, E., Dasgupta, S. & Thomas, G. (2004). ICT and business in the new economy: globalisation<br />

and attitudes towards ecommerce. Journal of global information management, 12(3): 44-64, July-September.<br />

Schmitt, P. and Michahelles, F. (2009). Status of RFID/EPC Adoption. Auto-ID Labs White Paper WP-BIZAPP-048.<br />

March 2009<br />

Sharma, A., Thomas, D., Konsynski, B. (2008). Strategic and Institutional Perspectives in the Evaluation, Adoption<br />

and Early Integration of Radio Frequency Identification (RFID): An Empirical Investigation of Current and<br />

Potential Adopters. Proceedings of the 41st Hawaii International <strong>Conference</strong> on System Sciences (HICSS),<br />

Hawaii, HI.<br />

Shih, D.-H., Chiu, Y.-W., Chang, S.-I., Yen, D.C. (2008). An empirical Study of Factors affecting RFID’s Adoption in<br />

Taiwan. Journal of Global Information Management, 16 (2), 58-80.<br />

Shulman, R. (2001), "Perishable systems take center stage", Supermarket Business, Vol. 56 No.4, pp.47-8.<br />

Roztocki, N. & Weistroffer, H. R. (2009). Information and Communications Technology in Developing, Emerging and<br />

Transition Economies: An Assessment of Research. AMCIS 2009 Proceedings. Paper 652.<br />

http://aisel.aisnet.org/amcis2009/652<br />

Strüker, J., Gille, D. (2008). The SME Way of Adopting RFID-Technology – Empirical Findings from a German crosssectoral<br />

Study. 1<strong>6th</strong> <strong>European</strong> <strong>Conference</strong> on Information Systems (ECIS 2008), Galway.<br />

The World Retail Congress. (2010). Day three programme. Daily Magazine, in association with Retail Week. October<br />

25-27, Berlin.<br />

Uhrich, F., Sandner, U., Resatsch, F., Leimeister, J.M. & Krcmar, H. (2008). RFID in retailing and Customer<br />

Relationship Management. Communications of the Association for Information Systems, 23: 219-234.<br />

United Nations. UNCTAD Secretariat. (2007). The information economy report 2007-2008: science and technology<br />

for development: the new paradigm of ICT. http://www.unctad.org/en/docs/sdteecb20071_en.pdf. [20June<br />

2008].<br />

Whitaker, J., Mithas, S., Krishnan, M.S. (2007). A Field Study of RFID Deployment and Return Expectations.<br />

Production and Operations Management, 16 (5), 599-612.<br />

Wyld, D.C., Jones, M.A., Totten, J.W. (2005). Where is my Suitcase? RFID and Airline Customer Service. Marketing<br />

Intelligence & Planning, 23 (4), 382-394.<br />

359


Towards a Comprehensive Evaluation Framework for ICT for<br />

Development Evaluation – an Analysis of Evaluation<br />

Frameworks<br />

Caroline Pade-Khene and Dave Sewry<br />

Rhodes University, Grahamstown, South Africa<br />

c.khene@ru.ac.za<br />

d.sewry@ru.ac.za<br />

Abstract: Information and communication technologies (ICT) can support rural development activities through<br />

providing supportive information and knowledge, and creating essential interconnectivities between rural areas and<br />

more developed regions. However, rural ICT for development (ICT4D) is still at best a ‘working hypothesis’, faced<br />

with barriers and challenges associated with implementation and use in the rural environment; which threaten the<br />

success, sustainability or relevance of ICT interventions in developing countries. The evaluation of rural ICT projects<br />

is indispensable as it determines the need, effectiveness, impact, sustainability and extent of the awareness of the<br />

contribution such projects or programmes can make in poverty alleviation and development. Even so, existing ICT for<br />

development evaluations are confronted with shortcomings and challenges which influence the reliability of<br />

evaluation conclusions. These shortcomings highlight the need to embark on a more comprehensive evaluation<br />

approach, sensitive to the rural environment. The paper adopts a theoretical multi-method approach to determine the<br />

multiple variables and components associated with rural ICT evaluation, and then to determine how these variables<br />

interrelate. The approach is founded on programme evaluation, ICT for development evaluation, and information<br />

systems evaluation. Programme evaluation informs this research study of the fundamental domains of evaluation<br />

and the structure that shapes an evaluation. An analysis of ICT for development evaluation frameworks and<br />

information systems frameworks compares and contrasts key characteristics that form the structure of a<br />

comprehensive evaluation that can be applied in ICT for development projects. The analysis and a review of<br />

programme evaluation enables the development of a Rural ICT Comprehensive Evaluation Framework (RICT-CEF)<br />

that encompasses the key components essential for a comprehensive evaluation of rural ICT projects. The<br />

theoretical framework aims to inform ICT intervention to improve and support rural development, through the<br />

application of fundamental and interconnected evaluation domains sensitive to the rural environment, throughout the<br />

project’s lifecycle.<br />

Keywords: ICT for development, information systems evaluation, rural development, programme evaluation<br />

1. Introduction<br />

Rural development is a significant focus for developing countries toward reaching their goals in<br />

development and poverty alleviation. ICTs are known to be key tools that have the potential to enhance<br />

rural development activities, as the availability of information and knowledge can support the operation<br />

and effectiveness of rural development activities. However, the growing establishment of ICT projects in<br />

rural areas signifies the need for accountability and an evaluation of the effectiveness, impact, and<br />

sustainability of such projects. There are cases where ICTs are known to have social and economic<br />

benefits in communities, but there are also cases where the implementation of ICT projects has actually<br />

not made a difference, or the effects have been harmful in communities (Tacchi, Slater and Lewis 2003).<br />

ICT4D actors need to understand that ICT4D is, at best, a “working hypothesis” where many key<br />

questions remain largely unanswered, with no concrete or credible data to support a wide range of claims<br />

concerning the use of ICT for development (Wagner, Day, James, Kozma, Miller, and Unwin 2005;<br />

Moodley 2005). Therefore, the evaluation of ICT4D initiatives is essential. Nevertheless, Pade and Sewry<br />

(2009) illustrate that current ICT4D evaluations are confronted with shortcomings and challenges which<br />

influence the reliability of evaluation conclusions. Embarking on a multi-method approach of evaluation<br />

theory can provide a foundation and guideline for developing a more comprehensive framework that<br />

addresses and considers the shortcomings and challenges of current ICT4D evaluation.This research<br />

paper introduces programme evaluation as a foundation for defining evaluation. A summary of key<br />

aspects that are considered across all the evaluation frameworks are explored as part of the process of<br />

constructing a comprehensive evaluation framework which is sensitive to rural ICT environments.<br />

Subsequently, a skeleton view of the Rural ICT Comprehensive Evaluation Framework (RICT-CEF) is<br />

presented. It is concluded that in order to obtain a better understanding and application of the RICT-CEF,<br />

a real-life case study investigation can reveal the lessons learned (shortcomings and suitability) from<br />

applying such a framework in a rural environment.<br />

360


Caroline Pade-Khene and Dave Sewry<br />

2. The foundation of programme evaluation<br />

Programme evaluation can be defined as (Rossi et al. 2004): “The use of social research methods to<br />

systematically investigate the effectiveness of social intervention programs in ways that are adapted to<br />

their political and organisational environments and are designed to inform social action to improve social<br />

conditions”. In contrast to basic research, the main objective of programme evaluation is not necessarily<br />

the discovery of new knowledge, but rather the “study of the effectiveness with which existing knowledge<br />

is used to inform and guide practical action” (Clarke 1999). This research study adopts Rossi et al.’s<br />

(2004) defined domains of programme evaluation, which include: Needs Assessment, Programme<br />

Theory Assessment, Process Assessment, Outcome Assessment, Impact Assessment, and Efficiency<br />

Assessment. The purpose and goals specific to a domain of evaluation attach a unique evaluation<br />

structure for each domain.<br />

3. A comprehensive evaluation<br />

Customarily, most researchers define the evaluation of a programme around assessing the impact it has<br />

on the target population. However, other aspects need to be considered when determining the<br />

effectiveness of a programme, which are related and build on each other to produce a more profound<br />

evaluation. Each evaluation domain applies a different approach to analyse an aspect of a programme,<br />

developing results that may build onto other evaluation types. In this case, a comprehensive evaluation,<br />

which is the ideal, becomes essential. Rossi et al. (2004) indicates that a comprehensive evaluation<br />

constitutes a process evaluation, which builds into an impact evaluation. On the other hand, Babbie and<br />

Mouton (2001) provide a more broad definition, where a comprehensive evaluation includes the<br />

application of all five domains of evaluation to assess the effectiveness of a programme. The<br />

development of an evaluation framework to apply to a particular programme intervention requires that<br />

there be a frame of reference (theory or model), that provides a base to understand the programme’s<br />

goals and context in which it operates (Babbie and Mouton 2001). Analytical frames of reference, such as<br />

theories and models, fulfil the following important functions: provide definitions of key concepts, guide<br />

data-collection and analysis, and provide an explanatory (causal) framework for the interpretation of<br />

empirical findings (Babbie and Mouton 2001). Consequently, the development of a comprehensive<br />

evaluation framework for rural ICT projects requires a collective analysis of the various theories, models,<br />

and aspects that may constitute the evaluation of a rural ICT project.<br />

4. An analysis of evaluation frameworks<br />

The factors associated with ICT project implementation in rural areas suggest the need to embark on a<br />

multi-method approach to determine the common multiple variables associated with rural ICT evaluation,<br />

and then to determine how these variables interrelate (Lannon 2007; Hudson 2001b). Two evaluation<br />

areas provide useful insight for the development of a comprehensive evaluation framework: the<br />

evaluation of ICT for development, and Information systems evaluation. ICT4D evaluation frameworks<br />

point out essential factors that should be built into evaluation based on the experience and lessons<br />

learned that have been applied to develop and design such frameworks. An analysis of IS evaluation<br />

frameworks also provide an outlook of aspects that can be applied in ICT4D projects to assess its value<br />

and usefulness, in volatile rural environments. A template based on foundational literature on programme<br />

evaluation, the need for ICT4D evaluation, and the shortcomings and challenges of current ICT4D<br />

evaluation was applied to each evaluation framework to provide a comparative analysis. The evaluation<br />

template produces a review of key frameworks that can contribute to the development of a<br />

comprehensive framework. The ICT4D evaluation frameworks selected are listed in Figure 1 and the IS<br />

evaluation frameworks selected are listed in Figure 2.<br />

A summary of an exploration of key aspects across the evaluation frameworks is discussed below. The<br />

review focuses on fundamental characteristics and approaches that can contribute to the development of<br />

a comprehensive evaluation framework for ICT4D evaluation.<br />

4.1 The project lifecycle stage at which the evaluation is conducted<br />

Lifecycle stages differ in five common ways:<br />

Pilot stage and/or proposal stage (frameworks 1 & 14): Essential to determine whether the ICTD pilot<br />

project is relevant or worth being taken to scale.<br />

361


Caroline Pade-Khene and Dave Sewry<br />

Annually (framework 2): A less time consuming and costly approach to evaluation, and allows a<br />

comparison to be made throughout the years of implementation and operation of the ICTD project.<br />

However, it is still important to regard the lifecycle stage<br />

Figure 1: List of ICT4D evaluation frameworks analysed<br />

Figure 2: List of IS evaluation frameworks analysed<br />

Only implementation and post-implementation stages (frameworks 6, 9, 12 & 13): Evaluation aimed<br />

at applying formative (process improvement) and summative (impact) evaluation. Lessons learned<br />

indicate that such frameworks expand their approach to evaluating other project stages. When ICT4D<br />

evaluation occurs during post-implementation, the aim is not to close the project. Rather, a project’s<br />

process/implementation will be assessed to learn lessons and improve or enhance the approach to<br />

implementing similar projects in other rural areas, once a successful or failed project has been fully<br />

implemented.<br />

Post implementation (framework 5): A typical stage conducted in ICTD evaluation. Impact<br />

assessment timing is crucial, and it is infeasible and unrealistic to assume that an ICT initiative can<br />

impact a community over a short period of time. A possible solution is to apply an iterative approach<br />

to impact assessment (as part of a comprehensive framework); identifying impacts that result and<br />

build up over time (Hudson 2001b; Pade and Sewry 2009).<br />

Throughout the life of the project (framework 3, 4, 7, 8, 14, & 15): Common evaluation domains<br />

associated with this approach include a needs assessment, process assessment and impact<br />

assessment across the lifecycle stages of the project. Some frameworks additionally consider<br />

efficiency assessment or programme theory assessment. Frameworks that appear to be or state they<br />

are applied throughout the life of the project, does not guarantee the evaluation is comprehensive. All<br />

domains of evaluation need to be applied throughout the project life cycle.<br />

4.2 The project environment<br />

The environments that most ICT projects target are rural, remote or marginalized areas in developing<br />

countries. An evaluation, like a project (Schwalbe 2006), cannot operate in isolation and needs to be<br />

sensitive to the complexities and factors of their environments. The project environment for an<br />

362


Caroline Pade-Khene and Dave Sewry<br />

information systems evaluation is usually not specified. However, in two particular frameworks (12 &15),<br />

the project environment is indicated, as the authors view evaluation as being partly shaped by the<br />

environment in which it is applied. Evaluation is seen as a continuous learning process as the operation<br />

and use of the system is sensitized to needs of its environment, through improvements applied and<br />

suggested from preceding evaluations (framework 12; framework 15).<br />

4.3 Domains of evaluation<br />

The evaluation domains (related to programme evaluation) include:<br />

Needs Assessment: Only four ICTD frameworks (frameworks 3, 4, 5 and 10) and two IS evaluation<br />

frameworks (frameworks 14 & 15). It aims to understand and set the desired priorities of the rural<br />

community and propose solutions that can be supported by ICT. The assessment should also<br />

accommodate demand driven needs that emerge over-time. IS evaluation suggests that a needs<br />

assessment should explore the project’s contribution and alignment with business strategy (rural<br />

development strategy). Furthermore, the assessment should be based on 1) the need for a<br />

development solution, and 2) an awareness of potentially new ICT initiatives relevant to the rural<br />

context.<br />

Programme Theory Assessment (framework 14): This domain specifically assesses whether the<br />

conception of an ICT4D project is actually designed to support rural development programmes.<br />

Where an assessment reveals the programme theory is flawed, responsible stakeholders need to<br />

redesign the programme. The domain is not applied or considered across the ICT4D evaluation<br />

frameworks and only in one IS framework; yet it is a crucial evaluation domain in such projects. The<br />

evaluation exercise described in framework 14 initially focuses on the project proposal stage, where a<br />

decision has to be made on whether to go ahead with the IS project or not. The types of theory<br />

assessed, relate to some of the tasks of the framework, e.g. implementability, risk, analysis of<br />

benefits, etc.<br />

Process Assessment: commonly applied across most of the evaluation frameworks analysed<br />

(frameworks 1, 2, 5, 7, 8 and 15). Aspects evaluated include project objectives, output and planning,<br />

sustainable ICT characteristics; all targeted at stakeholders involved in managing a project locally.<br />

Process assessment is generally seen as a procedure that results in suggested improvements to a<br />

project’s implementation, but most importantly, it should be seen as a continuous learning process (or<br />

iterative), given the complexity of rural environments.<br />

Impact and Outcome Assessment: a popular evaluation domain which all the evaluation frameworks<br />

apply. They are to an extent interchangeable, as most evaluation frameworks tend to describe them<br />

as similar aspects; however, Rossi et al. (2004), define them as separate domains. Some<br />

frameworks describe their specific outcome/impact assessment goals or approaches and provide<br />

useful concepts that should be included in the assessment of ICT for development projects, e.g.<br />

strengthening existing social and organisational community structures, gender sensitivity, etc. Impact<br />

assessment can be applied in two ways, i.e. 1) comparing before implementation and after<br />

implementation data, 2) assessment only post-implementation using retrospection, and ‘contrary to<br />

the fact questions (framework 8). Framework 14 sites that two elements need to be recognised in the<br />

assessment of outcome, that is, firstly the two stage benefits realization process (short-term<br />

(outcome) & long-term (impact)), and secondly, when developing impact theory, to make realistic<br />

assumptions concerning the project’s impact.<br />

Efficiency Assessment (frameworks 2, 11 and 14): Usually consists of a cost-benefits analysis and a<br />

cost-effectiveness analysis. This adds rigour to an impact assessment, and should be considered as<br />

part of a more comprehensive evaluation, rather than an isolated evaluation approach. A variety of<br />

costs for ICT4D can be determined, and the literature presents a variety of perspectives or<br />

approaches to financial assessment for cost benefit analysis (Heeks and Molla 2008).<br />

Additional evaluation domains identified include:<br />

A Baseline Study (frameworks 4 and 8): Serves to inform external stakeholders of the social, cultural,<br />

economic, political and technological (readiness) status of a community, as well as introduce<br />

community members to the project. It also acts as a foundation for a needs assessment and adds<br />

value to an impact assessment (comparing the before and after status of community).<br />

Flexibility Assessment (framework 14): An important domain in the context of rural ICT projects,<br />

whose environment is unpredictable at times (demand driven needs also emerge). Two aspects that<br />

should be considered in assessing flexibility include: 1) Examine the future changes that the ICT<br />

363


Caroline Pade-Khene and Dave Sewry<br />

project may be subjected to in the rural environment, and then ensure that these changes can be<br />

easily accommodated, and 2) an assessment needs to be done to determine whether the proposed<br />

ICT4D project is building in constraints to flexibility (for example, the project should be viewed<br />

holistically, considering how other traditional information flows (word of mouth, community meetings<br />

etc.) may be affected negatively or positively by the project).<br />

Scalability Assessment (framework 1): Determines whether a pilot project should be brought to scale.<br />

It assesses the conditions that are needed to replicate or scale up the project, and what might be the<br />

result of large scale application. ICT projects should begin at pilot level, to observe their success or<br />

potential sustainability.<br />

4.4 The evaluation model/approach/process<br />

The approach/model applied in each framework is characterised by the purpose of the evaluation, the<br />

project’s environment, and in some cases, the evaluation domain. The approaches applied can be<br />

categorized into:<br />

A process approach (frameworks 1, 3, 7, 10, 11, 12 and 13): This is the most common approach<br />

applied across the evaluation frameworks. This approach describes the step by step procedure to<br />

conduct an evaluation. It is not necessarily dependent on the lifecycle stage of a project, like an<br />

evaluation lifecycle approach, but rather focuses on the process or procedure to conduct an<br />

evaluation of any particular domain. In this case, a generic procedure is applied, which can be<br />

applied to assess various domains on which the evaluation framework intends to focus.<br />

A dimension focused approach (frameworks 2, 9 and 14). Evaluation tasks or elements that shape<br />

the process to conduct an evaluation are applied. There is not necessarily a standard sequence for<br />

the tasks/elements stated to carry out the evaluation, but some logic is required to determine an<br />

appropriate sequence to carry out the evaluation of particular projects. The evaluation tasks may<br />

therefore occur simultaneously, independently, or sequentially.<br />

An evaluation lifecycle approach (frameworks 4, 8 and 15): This approach applies the stages of an<br />

evaluation, based on the domains of evaluation. These domains are conducted in sequential order,<br />

depending on the lifecycle stage the project is on.<br />

A case study approach (framework 5 and 6): This evaluation process is a simplistic approach, made<br />

up of qualitative or quantitative methodologies and procedures used to collect the data, and the<br />

results reported. Other aspects of the frameworks (evaluation questions, data collection instruments,<br />

etc.) provide useful insight for ICT4D evaluation.<br />

4.5 Structure of the evaluation<br />

4.5.1 Evaluation questions and indicators<br />

The types of evaluation questions vary, depending on the purpose of the evaluation. Not all the<br />

frameworks specify their evaluation questions, specifically framework 11, 14 and 15. Here, evaluation<br />

questions can be proposed based on their evaluation goals. The questions from the frameworks provide<br />

a useful guideline for the kind of questions that can be asked under each evaluation domain, but are not<br />

necessarily the standard exhaustive list. Some questions may be generic and applicable to any ICT4D<br />

project, but some are for specific types of projects such as education, e-governance, etc. Therefore,<br />

generic questions identified under each domain of evaluation can be used across different ICT4D project<br />

evaluations.<br />

4.5.2 Nature of the evaluator-stakeholder relationship<br />

Of the evaluation frameworks analysed, there is an almost equal distribution of independent,<br />

participatory, and empowerment evaluation:<br />

Independent evaluation (frameworks 1, 2, 5, 6, 9 and 11). An independent evaluator is chosen to plan<br />

or design the evaluation, conduct the evaluation, and disseminate results; often stipulated by the<br />

evaluation sponsor. It is assumed that community (end-user) involvement is made adequate through<br />

the application of data collection methodologies to communicate their views, for instance, focus<br />

groups discussions and interviews. The nature of the evaluator-stakeholder relationship may depend<br />

on the domain of evaluation conducted, e.g. a cost-benefit analysis may need to be conducted by an<br />

evaluator experienced in financial modelling & analysis.<br />

364


Caroline Pade-Khene and Dave Sewry<br />

Participatory or Collaborative Evaluation (frameworks 4, 7, 8, 12, 13 and 15): The evaluator<br />

collaborates with participating stakeholders to plan, conduct and analyse the evaluation. Participatory<br />

research is applied, placing communities at the centre of the evaluation, through understanding the<br />

community’s existing social institutions, processes, and information ecology. For example, local<br />

stakeholders can participate in questionnaire design, specifying topics of local importance,<br />

suggesting culturally acceptable questions for ascertaining age and income levels, and identifying<br />

potential user groups to be included in the sampling frame and focus groups (framework 8). It aims to<br />

produce results for key stakeholders such as practitioners, funders and local participants or endusers.<br />

Empowerment Evaluation (frameworks 3, 8, 10 and 14): Occurs when participatory or collaborative<br />

evaluation additionally focuses on developing the capabilities of participatory stakeholders to conduct<br />

the evaluation. This contributes to an evaluation that is informed by local knowledge, develops local<br />

skill, encourages local ownership, builds demand for ICT services through the community incitement<br />

by local MCT staff, and promotes local sustainability (Scharffenberger 1999; Hudson 2001a). Target<br />

groups in a rural community, who have the potential to engage in and use ICTs to support their rural<br />

livelihood activities, can be selected and trained (empowered) to conduct the evaluation. The leading<br />

evaluator is not necessarily absent in conducting the evaluation when an empowerment approach is<br />

used, but plays a key role in supervising and guiding local stakeholders in the process.<br />

4.5.3 Methods and procedure applied<br />

Most frameworks apply a mixture of methodological approaches, that is, qualitative or quantitative, rather<br />

than focusing on one. Although there is no standard methodology or procedure across the frameworks,<br />

common methodologies can be identified which are assumed to be more effective for research in rural or<br />

marginalized environments. The mix and match of qualitative and quantitative approaches provides a<br />

deeper analysis of evaluation findings from different stakeholder or project perspectives. The Acacia<br />

Evaluation and Learning System framework (framework 8), suggests a useful procedure for mixing and<br />

matching different methodological approaches, called the ‘Rolling Design’ approach (action research),<br />

which integrates multiple data collection instruments that are sequentially implemented to obtain relevant<br />

data from a wide range of perspectives (Hudson 2001a). Examples of the various quantitative<br />

methodologies used include: questionnaires, quantitative data on technology, log sheets/records, and<br />

cost benefit analysis methodologies. Examples of the various qualitative methodologies used include:<br />

tracer studies, participatory appraisals, most significant change, interviews, photography, audio recording<br />

and video, focus groups, workshops, observation and participant observation, and document analysis.<br />

5. A comprehensive evaluation framework for rural ICT projects in developing<br />

countries<br />

An analysis of ICT4D frameworks and Information Systems frameworks, in addition to the concepts of<br />

programme evaluation, provides the fundamental elements necessary to shape and design a<br />

comprehensive evaluation framework for ICT4D projects. A high level (skeleton) illustration of the Rural<br />

ICT Comprehensive Evaluation Framework (RICT-CEF) is presented in this section.<br />

5.1 Defining comprehensive evaluation for rural ICT projects<br />

The comprehensive nature of a rural ICT evaluation is determined by the various domains applied to<br />

evaluate a project, as well as the key aspects that structure the evaluation process. Each evaluation<br />

domain applies a different approach to analyse an aspect of a project, developing results that may build<br />

onto other evaluation domains. Essentially, these different domains which are all key to the evaluation of<br />

a programme are interlinked, and where they are all applied interdependently, they contribute to a<br />

comprehensive evaluation of a rural ICT project (Babbie and Mouton 2001; Rossi, et al. 2004).<br />

Evaluation, in this case, is performed throughout the life of the rural ICT project, from Project Idea<br />

Generation to Post-implementation Review (in reference to the Rural ICT Project Life Cycle (RICT-PLC)<br />

proposed by Pade, Mallinson and Sewry (2008)). Therefore, comprehensive rural ICT evaluation is a<br />

continuous endeavour encapsulating different stages of ICT utilisation by target groups and individuals in<br />

rural communities. Comprehensive evaluation throughout the life of the project requires the evaluator to<br />

be sensitive to the rural environment, and especially aware of any social, political, cultural, or economic<br />

factors that can influence the evaluation form and scope at different stages of the rural ICT project.<br />

Accordingly, the evaluation approach should be multidisciplinary, comparative, participatory and<br />

multicultural to contribute to research on different elements associated with the contribution of ICTs to<br />

rural development. In this research, comprehensive rural ICT evaluation is therefore defined as: The<br />

365


Caroline Pade-Khene and Dave Sewry<br />

comprehensive application of all appropriate domains of evaluation throughout the life of a rural ICT<br />

project, in ways that are adapted to the social, political, cultural, and economic rural environment, and<br />

designed to inform ICT interventions to improve and support rural development.<br />

5.2 The structure of the Rural ICT Comprehensive Evaluation Framework (RICT-CEF)<br />

The RICT-CEF, depicted in Figure 3, is structured around three key components that guide the<br />

comprehensive evaluation of a rural ICT project or programme. The components include the overall<br />

evaluation plan, the comprehensive evaluation lifecycle, and the iterative design of the framework.<br />

Figure 3: The Rural ICT Comprehensive Evaluation Framework (RICT-CEF)<br />

5.2.1 The overall evaluation plan<br />

The Overall Evaluation Plan is applied to the overall evaluation process to guide the various evaluation<br />

domains. Providing a high level view of the evaluation plan, it comprises the key attributes that need to be<br />

366


Caroline Pade-Khene and Dave Sewry<br />

taken into account for planning the overall evaluation (considering the triple constraint of scope, time, and<br />

cost by Schwalbe (2006), and guidelines that should be applied across all the domains of evaluation<br />

(Clarke 1999; Hudson 2001a; APC 2008, Scharffenberger, 1999; Rossi et al. 2004; Wagner et al. 2005;<br />

Dymoke-Bradshaw and Cox 2003; Hedman and Borell 2005). However, each evaluation domain has its<br />

own local level evaluation plan that is specific to meeting the requirements and purpose of that particular<br />

domain. The overall evaluation plan provides a sketch and guideline for the whole comprehensive<br />

evaluation process, and acts as the initiation stage of the comprehensive evaluation framework.<br />

5.2.2 The evaluation lifecycle<br />

The Evaluation Lifecycle forms the central component of the framework, which describes the process and<br />

domains that a rural ICT comprehensive evaluation should go through. The lifecycle has domains that<br />

connect from the beginning to the end of the overall evaluation, and hence provide a gauge of the<br />

evaluation’s progress towards contributing to a comprehensive evaluation of the ICT intervention<br />

(Schwalbe 2006). The domains that make up the comprehensive evaluation cycle are:<br />

1. The Baseline Study: assesses the existing status (socio-economic and readiness) of the rural<br />

community. (frameworks 4 and 8)<br />

2. Needs Assessment: assesses the desired priorities and information needs for rural development,<br />

so as to propose appropriate solutions that can be supported by ICT. (frameworks 3, 4, 5, 6, 10, 13<br />

and 15)<br />

3. Programme Theory Assessment: assesses the conceptual design of the ICT project to support<br />

rural development programmes. (Rossi et al. 2004)<br />

4. Process Assessment: assesses how well the project is operating to implement its intended<br />

functions, as stipulated in the project plan. It also assesses the flexibility of the project. (frameworks<br />

1, 3, 4, 7, 8, 14 and 15)<br />

5. Outcome and Impact Assessment: assesses the intended and unintended outcomes or effects of<br />

the ICT project on the conditions of the targeted population. (frameworks 1, 2, 3, 4, 5, 10 and 14)<br />

6. Efficiency Assessment: assesses ICT intervention costs associated with project effects or impact.<br />

(frameworks 1 and 11)<br />

7. Scalability Assessment: assesses whether a rural ICT pilot project should be taken to scale.<br />

Therefore, this domain is only applied to ICT pilot projects. (frameworks 1 and 12)<br />

Each domain is described according to the following structure:<br />

a).The Purpose of the evaluation<br />

b).The Domain Process: The domain process either adapts one particular process from the<br />

frameworks analysed, or a combination of key processes and practices from different frameworks,<br />

that suit the requirements and purpose of the domain.<br />

c).The Evaluation Plan: Guides the evaluation domain process. It is tailored around three features,<br />

adopted from the evaluation structure of Rossi et al., (2004): 1) The Evaluation Questions, 2) The<br />

Methods and Procedures, and 3) The Nature of the Evaluator-Stakeholder Relationship. The<br />

framework provides examples of commonly used questions or methods that have been applied in<br />

other frameworks or ICT projects, which an evaluator may possibly choose from or adapt for other<br />

ICT4D project evaluations. The nature of the evaluator-stakeholder relationship, conversely, is the<br />

recommended approach for the particular evaluation domain.<br />

d) Pilot Project Aspects: Highlights additional aspects or factors that should be considered when<br />

evaluating a pilot project.<br />

e) Expected Outputs: The expected outputs of the particular domain.<br />

f) The Dissemination of Results: Specifies the target audience for evaluation results and the<br />

approach for documenting and presenting the results.<br />

g) Application Stage of the RICT-PLC: The evaluation framework makes use of the RICT-PLC<br />

proposed by Pade, Mallinson and Sewry (2008) to associate the different evaluation domains with the<br />

rural ICT project phases that may directly use the results of the different evaluations. Every ICT4D<br />

project lifecycle is unique. Even though various ICT projects may have different lifecycles or<br />

associated phases, the RICT-PLC provides a general picture for stakeholders of various ICT projects<br />

interventions to determine where best to apply a particular evaluation domain within their projects.<br />

367


Caroline Pade-Khene and Dave Sewry<br />

Nevertheless, the comprehensive evaluation of the project may initiate during the ongoing<br />

implementation of the project, and not from the inception of the project. In this case, the different<br />

evaluation domain results still serve to improve different aspects of the project that can be revised.<br />

h) Domain Dependence: Each domain of evaluation is interlinked to preceding or subsequent<br />

domains within the evaluation lifecycle (see Figure 3). This section explains the interdependent<br />

relationship between the domains.<br />

5.2.3 The iterative design of the framework<br />

This Iterative Design reflects the iterative nature of rural ICT projects, which adapt over time due to the<br />

progressive integration of technology in rural livelihoods, and the changing nature of the rural<br />

environment, as a result of development activities. In this case, particular evaluation domains are<br />

repeated over-time, building and developing the scope of the evaluation as the project’s intervention and<br />

impact progresses in the rural community (frameworks 12, 15 and 8). This is particularly essential to<br />

address the time-dependent nature of rural ICT projects, as many changes or impacts brought about by<br />

ICT initiatives are long-term and often indirect. ICT4D initiatives are a form of human development<br />

infrastructure which is highly time-dependent. The comprehensive evaluation lifecycle should be<br />

incorporated into the rural ICT project lifecycle, where evaluation results continuously contribute to<br />

improving and guiding the iterative (progressive) development of the ICT project. Within the iterative<br />

design of the evaluation framework, evaluators should document changes and identify evaluation aspects<br />

(needs, process (service utilisation), impact, etc.), that result over time, and also consider barriers that<br />

could hinder long-term impacts. Each iteration (iteration0, iteration1…) is initiated and driven by the<br />

changing needs of the community. The subsequent domains of evaluation are therefore influenced by<br />

changing needs, as the results of the needs assessment build into the evaluation of other aspects of the<br />

rural ICT project, such as, the process, outcome, impact and efficiency (see Figure 3).<br />

Iteration0 is associated with assessing a pilot project, and hence the scalability of the project to determine<br />

whether or not it should be taken to scale. The subsequent iterations (which comprise of only needs<br />

assessment through to efficiency assessment) contribute to evaluating the progressive development of<br />

the rural ICT project over-time. The duration or life of the ICT intervention is not always obvious, and in a<br />

number of cases, they are not sustainable in developing countries. Therefore, the iterative evaluation<br />

design may be appropriate for sustainable on-going ICT projects, but in cases where the ICT services are<br />

not being developed further or new ICT initiatives are no longer being introduced to the rural community,<br />

the evaluation design changes. In this case, the evaluation design in the long-term is centred only on<br />

assessing the impact of the project over time.<br />

6. Conclusion<br />

The evaluation of ICT4D projects is essential to determine their value in supporting a change in social<br />

conditions, and identifying ways to enhance or improve the project. Programme evaluation provides a<br />

foundation to understand the concept of evaluation for social development programmes. The types of<br />

domains of evaluation vary at different stages of a programme’s life cycle. An evaluation framework that<br />

applies all five domains in the evaluation of a programme is considered a comprehensive evaluation<br />

framework. Embarking on a multi-method evaluation approach is essential to develop an appropriate<br />

framework which is sensitive to the various factors associated with rural ICT project implementation. The<br />

results of the framework analysis and a review of programme evaluation highlighted the fundamental<br />

elements necessary to shape and design a comprehensive evaluation framework for ICT4D projects. A<br />

skeleton presentation of the RICT-CEF illustrates the overall evaluation plan, evaluation lifecycle, and<br />

iterative design fundamental for the comprehensive evaluation of rural ICT projects in developing<br />

countries. In order to obtain a better understanding and application of the RICT-CEF, a real-life case<br />

study investigation can reveal the lessons learned (shortcomings and suitability) from applying such a<br />

framework in a rural environment.<br />

References<br />

Association for Progressive Communications, The (2008) “Gender Evaluation Methodology”. [online],<br />

http://www.apcwomen.org/gem/?q=about.<br />

Babbie, E and Mouton, J. (2001). The Practice of Social Research, Oxford University Press, Cape Town.<br />

Batchelor, S. and Norrish, P. (2006). Framework for the Assessment of ICT Pilot Projects, The World Bank, InfoDev,<br />

Washington DC.<br />

Carlsson, S. A. (2003) “Advancing Information Systems Evaluation (Research): A Critical Realist Approach”,<br />

Electronic Journal of Information Systems Evaluation, Vol 6, No.2, pp 11-20.<br />

368


Caroline Pade-Khene and Dave Sewry<br />

Clarke, A. (1999) Evaluation Research: An Introduction to Principles, Methods and Practice (1e), London: SAGE<br />

Publications.<br />

Dymoke-Bradshaw, C. and Cox, B. (2003) “Developing Organisation-technology Fit: The Role of Evaluation in IS<br />

Innovation Projects”, Electronic Journal of Information Systems Evaluation, Vol 2, No.1.<br />

Fitzerald, G. (1998) “Evaluating Information Systems Projects: A Multidimensional Approach”, Journal of Information<br />

Technology, Vol 13, pp 15-27<br />

Gigler, B. (2004) “Including the Excluded – Can ICTs Empower Poor Communities? Towards an Alternative<br />

Evaluation Framework Based on the Capability Approach”, Paper read at 4 th International <strong>Conference</strong> on the<br />

Capability Approach. University of Pavia, Italy, September.<br />

Harris, R. W. (2001) “Using Stories for Evaluating Community Based Information Communication Technologies in<br />

Developing Countries”, Part of the report of the PANAsia Telecentre Learning & Evaluation Group (PANTLEG),<br />

[online], http://rogharris.org/Using_Stories.pdf.<br />

Hedman, J. and Borell, A. (2005) “Broadening Information Systems Evaluation through Narratives”, Electronic<br />

Journal of Information Systems Evaluation, Vol 8, No.2, pp 115-122.<br />

Heeks, R. and Molla, A. (2008) “Compendium on Impact Assessment of ICT-for-Development Projects”, Working<br />

Paper Series, No. 36, Institute for Development Policy and Management, Manchester.<br />

Hudson, H. E (2001a) The Acacia Programme: Developing Evaluation and Learning Systems for African Telecentres.<br />

in: Latchem, C. and Walker, D. (eds). Telecentres: Case Studies and Key Issues, The Commonwealth of<br />

Learning, Vancouver.<br />

Hudson, H. E (2001b) Telecentre Evaluation: Issues and Strategies. in: Latchem, C. and Walker, D. (eds).<br />

Telecentres: Case Studies and Key Issues. The Commonwealth of Learning, Vancouver.<br />

Kozma, R., McGhee, R., Quellmalz, E. and Zalles, D. (2004) “Closing the Digital Divide: Evaluation of the World<br />

Links Program”, International Journal or Educational Development, Vol 24, pp 361-381.<br />

Lannon, J. (2007) “Evaluating Online Human Rights Resource Centres”, Paper read at 16 th International <strong>Conference</strong><br />

on Information Systems Development (ISD2007),Galway, Ireland, August.<br />

Macome, E. and Cumbana, C (2002) “Evaluation Study of the Manhica and Namaacha Telecentres after 3 years of<br />

operations”, [online], Eduardo Mondlane University Informatics Centre,<br />

http://portal.unesco.org/ci/en/files/13411/10679419181Needs_assessment_MOZ.pdf/Needs%2Bassessment%2<br />

BMOZ.pdf.<br />

Moodley, S (2005) “The Promise of E-Development? A Critical Assessment of the State of ICT for Poverty Reduction<br />

Discourse in South Africa”, Perspectives on Global Development, Vol 4, No.1.<br />

Pade, C. and Sewry, D. (2009) “The Practice and Need for Rural ICT for Development Evaluation: An Experience of<br />

the Siyakhula Living Lab Baseline Study”, Paper read at 3 rd International Development Informatics Association<br />

(IDIA) <strong>Conference</strong>. Digitally Empowering Communities: Learning from Development Informatics Practice.<br />

Kruger National Park, South Africa, October.<br />

Pade, C., Mallinson, B. and Sewry, D. (2008) “Project Management Practice for Rural ICT Project Sustainability in<br />

Developing Countries”, Paper read at 2 nd <strong>European</strong> <strong>Conference</strong> on Information Management and Evaluation<br />

(ECIME), University of London, Royal Holloway, UK, September.<br />

Rossi, P. H., Lipsey M. W. and Freeman, H. E. (2004) Evaluation: A Systematic Approach (7e), Sage Publications,<br />

London.<br />

Scharffenberger, G. (1999) Telecentre Evaluation Methods and Instruments: What Works and Why? In Gomez, R.<br />

and Hunt, P. (eds), Telecentre Evaluation: A Global Perspective. International Development Research Centre<br />

(IDRC), Ottawa.<br />

Schwalbe, K. (2006) Information Technology Project Management (4e), Thomson Course Technology, Boston,<br />

Massachusetts.<br />

Tacchi, J., Slater, D. and Lewis, P. (2003). Evaluating Community Based Media Initiatives: An Ethnographic Action<br />

Research Approach. Paper read at the Information Technology for Development <strong>Conference</strong>, Oxford, United<br />

Kingdom, July.<br />

Wagner, D. A., Day, B., James, T., Kozma, R. B., Miller, J. and Unwin, T. (2005) Monitoring and Evaluation of ICT in<br />

Education Projects: A Handbook for Developing Countries, Washington DC: The World Bank (InfoDev).<br />

369


User Adoption of the New Health Information System -<br />

Individual and Organizational Determinants<br />

Bahlol Rahimi<br />

Urmia University of Medical Sciences, Iran<br />

bahlol.rahimi@umsu.ac.ir<br />

Abstract: The computer has become an integral part of healthcare delivery by providing tools for improvements in<br />

the quality of care. However, the failure rate for information systems implementation in healthcare organizations will<br />

pose a great challenge in the area. In this paper, we aim to discuss user adoption of the new health information<br />

system in healthcare organization. In this study, we review the literature in order to illustrate previous studies that<br />

present failures in implementing new IT-based systems in healthcare settings. Next, we discuss and apply theory to<br />

interpret the results. The results demonstrate some important aspects in the implementation of new health<br />

information system such as considering user involvement in the design and implementation phase of the system,<br />

taking into consideration and differentiating the different professional needs when developing the system, user<br />

training methods, and to consider work routine changes due to implementation of the new system. Our findings show<br />

that evaluations of health IT-based systems seem to have not been successfully implemented in a number of cases,<br />

and have failed to demonstrate improvements in patient care, operating costs savings, and improvements in<br />

productivity. This study shows that, in fact, we have not learned from the past. It can be concluded that the<br />

application of a participatory design approach in health IT-based systems may be fruitful for the organization. The<br />

involvement of the end users in the design and implementation process will allow for the collection of their ideas, thus<br />

benefiting the developing system.<br />

Keywords: user adoption, health information system, user involvement, participatory design<br />

1. Introduction<br />

The computer has become an integral part of healthcare delivery by accelerating organizational and<br />

provider-patient communication as well as providing tools for improvements in the quality of care (Rahimi<br />

& Vimarlund 2007). However, the failure rate for information systems implementation in healthcare<br />

organizations will pose a great challenge in the area (Kaplan 2000). This is in part because the big efforts<br />

that policy and decision makers do for implementation of those system such as huge investment (Poon et<br />

al 2004). In particular, the majority of attempts to introduce information systems initiatives into both the<br />

private and public sectors, including healthcare information systems that require data entry by healthcare<br />

practitioners, have been unsuccessful (Van Der Meijden et al 2003; Adi 2000).<br />

However, it cannot be ignored that a characteristic of many public sector reforms is that they are<br />

introduced into organizations that are not willing to accept the reform, have the technical ability to<br />

understand and implement the change, or possess the ability to maintain the change once it has been<br />

introduced. As a result, the reforms are often severely delayed or create distortions that have damaging<br />

effects, leading to changes that are often ultimately abandoned (Diamond & Khemani 2005).<br />

In light of the abundant experience available, it is difficult to understand why the frequency of<br />

implementation failures is so high. For example, it has been determined that half of all computerized<br />

information systems fail largely due to user resistance (Anderson 1997). Likewise, Paré et al (1998)<br />

showed that many healthcare organizations have devoted significant amounts of money and frustrated<br />

countless people in wasted efforts to introduce new information systems.<br />

These failures are alarming, since the introduction of new, large healthcare information systems require a<br />

major investment (Adi 2005). In order to have successful and sustainable information technology in a<br />

healthcare setting, many factors need to be considered when introducing the new technology. One factor<br />

that has been indicated as crucial for success is the involvement of clinicians and other staff who will use<br />

the healthcare information systems throughout the planning and preparation stages of the design<br />

(Sengstack & Gugerty 2004; Birkmann et al 2006; Nykänen & Karimaa 2006). In general, regular and<br />

open communication between designers and users allows for the understanding of user preferences,<br />

their needs, their problems, and confusions. Recently, participatory design has been employed in<br />

different fields (Joyes & Chen 2007); however, as for health informatics, which is a more recently<br />

developed field, the use of participatory design has a shorter history (Pilemalm & Timpka 2007). Due to<br />

obvious problems in introducing integrated computer-based health applications, utilizing new approaches<br />

such as participatory design can make a difference where traditional methods have failed, in particular,<br />

with regards to the social factors that are of importance for the successful implementation of an<br />

370


Bahlol Rahimi<br />

information system (Pilemalm & Timpka 2007). Participatory design, which explicitly advocates active<br />

user participation throughout the design process, is a proactive design method.<br />

In this paper we aim to discuss the underlying factors, mostly individual and organizational, for user<br />

adoption of new health information systems in healthcare organization.<br />

2. Theoretical background<br />

2.1 Participatory design<br />

The participatory design approach was introduced in Scandinavia in the 1970s (Schuler & Namioka<br />

1993). It started in the field of computer software design, but similar approaches have been used in other<br />

fields as well (Sjöberg & Timpka 1998). Various studies have found that involving users in the system<br />

development process is an effective way to increase their adoption of the system (Faber 2003). In this<br />

respect, participatory design can be an appropriate approach for promoting user involvement.<br />

Participatory design was initiated as a resource to empowering system users by involving them in the<br />

design of an important part of their work setting. By involving of healthcare staff in the design and<br />

development processes of health information systems, it is said that the usability of the systems as well<br />

as their success increase (Van der Meijden et al 2001). Furthermore, it is important to consider the<br />

attitude of end-users -instead of one single viewpoint on a system’s requirements- in the process of<br />

development in order to increase final acceptance (Vimarlund & Timpka 1998).<br />

In addition, users are able to redesign and evaluate their work routines by applying the experience<br />

obtained during the participation process, and are even provided with the opportunity to improve their<br />

understanding of computers (Vimarlund & Timpka 1998). However, another objective for participating in<br />

the process of participatory design is based on recognizing the fact that it gives designers new and better<br />

ways of gaining an understanding of the users’ everyday working practices. Additionally, users are<br />

expected to be more willing to accept the final system once it is introduced, when they have assisted<br />

developers to arrive at a more accurate and realistic model, and reduces their aversion to the new<br />

system. The final overall effect of this participation can be considered to be the main incentive for<br />

improving work efficiency and productivity. It is argued that participation helps users to enhance their<br />

skills, thereby increasing the quality of services provided.<br />

To have an effective and credible health information system, user involvement in the design and<br />

development processes influences the perception, reception, and ultimately, the success of the new<br />

system (Wild et al 2004).<br />

2.2 Social capital<br />

The political scientist Robert D. Putman (1993) explained social capital as a theory about the<br />

complexities of trust and participation (Putman 1993a). The concept refers to features of social<br />

organization such as norms and networks that can improve the efficiency of society by facilitating<br />

coordinated actions (Putman 1993b). The social capital vocabulary has expanded, and today it<br />

addresses broadly the value of social networks both between similar and different groups of people by<br />

norms of reciprocity. The narrowest concept of social capital however, is still associated with Putman<br />

(Putman 1993a, 1993b). He views it as a set of “horizontal associations between people”, and argues<br />

that social capital encourages the formation of social networks and associated norms that have an effect<br />

on the productivity of the community and it facilitate the coordinated cooperation for the mutual benefit of<br />

the members of this “association” (Putman 1993a). Prior to Putman’s study, the concept of social capital<br />

had been used in community studies to describe relational resources embedded in personal ties within a<br />

community (Cohen & Prusak 2001). Nowadays, it is being applied in a wide range of studies in different<br />

disciplines, from anthropology and economy to information systems. However, despite the variety of<br />

definitions available for social capital, there is a consensus about the fact that social capital consists of<br />

those specific processes among people and organizations, which arise from working collaboratively in an<br />

atmosphere of trust, leading to the accomplishment of a goal with mutually shared benefits (Putman<br />

1993b). The benefits of social capital have, however, been identified and considered as “optimal” when<br />

social capital is created by individuals who learn to trust one another so that they are able to make<br />

credible commitments and rely on generalized forms of reciprocity, contributing to an increase in society’s<br />

productive potential (Ostrom 2000).<br />

371


Bahlol Rahimi<br />

To summarize, social capital consists of individuals forming and utilizing social networks in communities.<br />

Like financial capital, social capital is a resource that helps to sustain a community and encourages the<br />

collaboration and cooperation between members of groups for mutual benefits (Preece 2002).<br />

Introducing a new, healthcare information systems in which all users are involved and where networks<br />

are used for collaboration at different levels of the healthcare setting environment, are examples of such<br />

assets. Furthermore, the networks are usually groupings of people with common interests.<br />

2.3 Human capital<br />

Human capital is defined as “the skills, knowledge, and experience possessed by an individual or<br />

population, viewed in terms of their value or cost to an organization or country” (McKean 2005) and “the<br />

skills, general or specific, acquired by an individual in the course of training and work experience”<br />

(Smullen & Hand 2005). The main components of human capital, as stated in the literature include<br />

qualifications and knowledge acquired through formal education as well as skills, competencies, and<br />

expertise acquired through on-the-job training. Several authors have recognized the importance of<br />

human capital as a firm’s only appreciable asset due to the fact that the individuals’ productivity improves<br />

with literacy. It is said that workers may increase their productivity at the workplace by learning new skills<br />

and improving old ones (Vimarlund & Timpka 1998). The introduction of new technologies is likely to<br />

require the employment of a highly educated and skilled staff. There is, however, substantial evidence<br />

that education and training have strong positive effects on the accumulation of physical capital (Blundell<br />

et al 1999). In the case of implementing an healthcare information systems, acquiring knowledge and<br />

experience through formal and informal training becomes important and any shortfalls in the program<br />

raise many problems such as users’ complaints and resistance against the system.<br />

3. Methods<br />

In this study, I first review the literature in order to illustrate previous studies that present failures in<br />

implementing new IT-based systems in healthcare settings. Next, I discuss and apply the concepts just<br />

discussed to interpret the results.<br />

A literature review of published evaluation studies of IT-based systems in healthcare during the last five<br />

years (from 2003 to 2007) was performed. Linköping University’s database was used to gain access to<br />

articles on this subject. The keywords used in the search were: patient records, medical records, health<br />

records, information technology, medical informatics, healthcare information, health informatics, hospital<br />

information system, patient care information system, information system/technology, and evaluation<br />

study. I included papers that demonstrated a failure in implementing IT-based systems as well as some<br />

studies that discussed success and failure factors affecting the systems.<br />

4. Context of the study<br />

Study context of the papers including in this study has been mostly the Östergötland County council in<br />

Sweden, where tax-financed healthcare services are provided to the residents by the county council.<br />

Sweden has a decentralized healthcare system, with 20 county councils and 290 municipal councils as<br />

principals and care providers. Their responsibility as principals includes the provision of adequate care<br />

services and the requirement to develop, finance, and assure quality of all care activities. Implementation<br />

of a new integrated computerized patient record system was initiated in 2007 as a pilot project at a<br />

healthcare centre in the west part of the county, Motala. The implementation process continued from the<br />

west part (Motala) to the east part (Norrköping) of the county and was finished by the end of 2008. This<br />

new integrated system, developed commercially, provides a comprehensive overview of the patient’s<br />

health conditions and care. The system provides an infrastructure for sharing patient data between all<br />

healthcare care providers within the county council.<br />

5. Results<br />

A literature review was performed for evaluation studies of IT-based systems in healthcare. In our review<br />

we found that during the period 2003– 2008, most of the evaluation studies aimed to include issues such<br />

as the effectiveness of the systems, the quality of care, user and patient satisfaction, and the system’s<br />

usability. Based on the review literature, introduction and use of the computerized patient record systems<br />

was found to have positive effects such as economic benefits, high acceptance score and satisfaction<br />

among the users in the implemented sites and also improvements in management and work process.<br />

Moreover, introducing the telemedicine systems was found to have positive effects such as reducing<br />

spent time per patient during the visiting by clinical staff, economic benefits, and also quality of care.<br />

Regarding to the decision support system’s studies, it can be seen that introducing a clinical decision<br />

372


Bahlol Rahimi<br />

support systems in healthcare organizations had positive effects such as improved quality of care,<br />

satisfaction among users in the implemented sites, and also improvements in management and work<br />

process.<br />

The data captured in this study to show importance of evidence-based information system<br />

implementation in healthcare, an explorative case study design based on a single case was used for data<br />

collection and analysis. In this study the data were collected during a period of four months through<br />

interviews and document analyses. In the first step, representatives from all professional categories<br />

(physician, nurse, social worker, administrator, and pharmacist) using the new integrated computerized<br />

patient record system were interviewed by a member of the studied county council. Then 34 interviews<br />

conducted to validate our first results. Each interview session lasted about two hours. All documents<br />

published by the county council and the local magazines and newspapers that mentioned the system<br />

were also reviewed.<br />

In this study, different actors’ perspectives were considered when collecting data and analyzing them. We<br />

have used physicians, pharmacist, and nurses and other healthcare staff’s perspectives for our analysis.<br />

Considering different actors’ perspectives are needed to study in order to understand the likely<br />

accomplishment of new ICT goals in which it built in the area of health informatics. We categorized our<br />

finding into three groups that are: informatics skills, design interaction, and expectation.<br />

5.1 Informatics skills<br />

The data showed the nurses and other non-physician staffs were particularly unsatisfied, because they<br />

felt that the training sessions were based mostly on physicians’ needs. It was found that a failure to give<br />

all groups of users’ adequate training in using the integrated computerized patient record system<br />

negatively impacted the outcome of the implementation process. For instance, because the nurses had<br />

not learned to use the system functions properly, they found that the new practice routine was time<br />

consuming. Once the system was implemented, ongoing support was reported to be crucial for the<br />

success of the newly implemented system. They asked for the option of further training in order to<br />

overcome day-to-day problems.<br />

5.2 Interaction design<br />

The first technical problem was that logging on to the integrated system was perceived as consuming too<br />

much time. Then, after logging on to the system, several functions were found to be unintuitive and not<br />

user-friendly, causing dissatisfaction and disappointment. With the new system, calling up a specific file<br />

consumed more time than the previous system. The integrated system also required use of new terms<br />

and concepts, and the users emphasized that learning these took time.<br />

5.3 Expectation<br />

In the case study setting, the users expressed that more user participation in the design and<br />

implementation phase of the system would have provided a better fit into workflows and work practices.<br />

There was a general unwillingness to adapt clinical routines to the new system. The main adjustment of<br />

the implementation process that the users – especially physicians – asked for was “more involvement in<br />

the decision procedures”. More user involvement would both have helped define the system<br />

requirements in more detail and revise work practices to better integrate the new system. The<br />

respondents also made complaints about the timing of the implementation at the pilot site. They felt that<br />

the policy-makers had decided to implement the system in too short a time period, causing problems with<br />

adjustments, mainly in learning terms and navigation routines.<br />

5.4 What do end-users say is missing?<br />

The consequences of this trial seem to indicate that medical informaticians must learn from the past in<br />

order to communicate and adapt their own practices according to experience and research. The results<br />

demonstrate the lack of some important aspects in the implementation of a new system:<br />

1. User involvement in the design and implementation phase of the system would provide better<br />

insight into existing workflows and work practices. From the users’ point of view, to better integrate<br />

the new system, involving users in the system design and implementation phase would have avoided<br />

some human-computer interaction problems as well as training problems.<br />

373


Bahlol Rahimi<br />

2. In fact, perhaps the most important question is concerned with the way in which the<br />

implementation process can be adapted to different professional needs. It is clear that different staff<br />

members in the hierarchy of a healthcare organization demand different interfaces and modules. For<br />

example, taking into consideration and differentiating the nurses’ needs from the physicians’ needs,<br />

when developing the system, become important in order to achieve a good adoption rate for the new<br />

system. To understand how the implementation of a new system can be adapted to a large variety of<br />

user needs and expectations seems to be an important step towards facilitating user adoption of<br />

ICPRs.<br />

3. Staff complained that the training provided was not based on their need.<br />

4. The changes made to the current work routine due to the new system were also a source of<br />

resistance against the adoption of the system.<br />

5. Additionally, the results show that one of the important factors influencing the outcome of health ITbased<br />

systems implementation is the design of the user-training programs. Users’ inability to<br />

effectively interact with the new system caused secondary issues, leading to opinion that the health<br />

IT-based systems were not appropriately designed.<br />

6. Discussion<br />

In this study we found that, most of the evaluation studies aimed to include issues such as the<br />

effectiveness of the systems, the quality of care, user and patient satisfaction, and the system’s usability.<br />

A key requirement for analytical work is a clear definition of what constitutes an IT-based application.<br />

Today there are different terminologies relating to similar systems. Analysis of the effects has<br />

predominantly taken place at an organizational level. However, this does not take account of the<br />

differences in the settings in which analysis is taking place.<br />

It has been difficult to find generalized models and methods to evaluate IT-based applications in clinical<br />

settings that cover all such aspects as economic and both inter-organizational and intra-organizational<br />

approaches. The review reported in this study found that no standard framework exists for evaluating or<br />

developing evaluations and obtaining clearer and more exact feedback about the implemented systems’<br />

effects, or about the outputs of implementation and the use of ICT in healthcare settings. It also found<br />

that no previous studies have explored the impact of ICT on the healthcare systems’ productivity and<br />

effectiveness.<br />

A significant barrier to investment in ICT in healthcare is the widely recognized fact that any cost saving<br />

resulting from technology changes is not always seen by the implementer, but is rather passed on to a<br />

third party. In essence, benefits appear at one site and in one budget, while a large share of the cost<br />

commitments appear at another site and in another budget. To our best knowledge, the evaluation<br />

studies performed do not include any discussion about this important issue, or how lack of incentive to<br />

adopt systems can influence the organization and its personnel. Further, the potential effects of the<br />

implementation of IT-based applications are identified without analyzing them from an inter-organizational<br />

and economic perspective (Nykänen & Karimaa 2006).<br />

The case study comparisons with the literature review found both a recurrence of previously reported<br />

implementation problems and the development of new issues specific to the integrated system context.<br />

Possibly, the most important challenge in the case study setting concerned the way in which the<br />

integrated computerized patient record system implementation process could be adapted to the needs of<br />

different user groups. The second important challenge associated with it was the human-computer<br />

interaction consequences of the large-scale technical integration of sub-systems into a homogenous<br />

infrastructure. In this study re-experiences of known implementation problems were found. The fact that<br />

users’ training were based on physicians’ needs and not adjusted to nurses and other non-clinicians was<br />

a major source of complaints. From the case study setting, requests were expressed for user involvement<br />

in the design and implementation phase of the system, in order to provide better insights into existing<br />

workflows and work practices (Joyes & Chen 2007).<br />

A HIS system with integrated clinical decision support can be an advantage for the busy clinician who<br />

must combine and manage an increasing body of clinical knowledge. However, such support will not be<br />

optimal if clinicians begin to trust these systems without questioning the assistance (Pilemalm & Timpka<br />

2007). Recent research on safety in man-machine interaction suggests that the presence of<br />

environmental cues reflecting hazards increases alertness among decision-makers and reduces the risk<br />

of mistakes. From this perspective, it is positive that the system users had doubts about the reliability and<br />

374


Bahlol Rahimi<br />

completeness of the support provided by the system. Alertness is particularly important in light of the<br />

results of several previous studies that reported that HIS systems led to a number of errors and adverse<br />

drug events.<br />

Our findings show that evaluations of health IT-based systems seem to have not been successfully<br />

implemented in a number of cases, and have failed to demonstrate improvements in patient care,<br />

operating costs savings, and improvements in productivity. This study shows that, in fact, we have not<br />

learned from the past. Weng et al. (2006), in their paper aiming to apply participatory design for the<br />

creation of a highly usable collaborative protocol writing system for a national cancer clinical trial protocol<br />

authoring organization, show that after applying participatory design, users liked the system design. The<br />

users also liked the fact that some features of the system could speed up the reviewing process, the<br />

interoperability between the new system and the old tools, the integration, and the “all-in-one” design.<br />

The system’s usability was demonstrated in their results. Their study concluded that they could reduce<br />

usability errors of healthcare information systems with concrete approaches to participatory design.<br />

Therefore, before draining further the scarce resources available to healthcare organizations, it is<br />

necessary to develop strategies that will involve the end-users of the new system and ask for their<br />

assistance in accelerating the organizational change as well as work routines. Vimarlund et al (2001), in<br />

their paper discuss the economic impacts from the use of participatory design for the development of<br />

public-health information systems. They argue for the need of a method that minimizes the information<br />

asymmetry in the development process, thus avoiding market failures. They conclude that participatory<br />

design is beneficial for the public-health organizations involved in information system development<br />

processes by avoiding the risk of rejection of the final system, while increasing knowledge, decreasing<br />

information asymmetry, updating habits and work routines in a “natural manner”, avoiding unnecessary<br />

sunk-costs, and contributing to the efficient use of human and economic resources (Weng et al 2007).<br />

The challenges we have described first and foremost highlight a lack of collaborative approaches in the<br />

implementation of HISs. For instance, many of the most significant failures emerge from the absence of<br />

feedback from end users to developers during the development process. The logical conclusion is to<br />

increase the involvement of end users in the implementation process in order to provide:<br />

Designers with new and better ways of gaining an understanding of users’ everyday work practices.<br />

Users with the opportunity to redesign and evaluate their work routines by applying experience<br />

obtained during the participation process<br />

Users with the opportunity to appreciate the HIS as a useful tool in their everyday working life<br />

In fact, what is needed is the use of an implementation methodology that minimizes the information<br />

asymmetry in the implementation process, and allows the accumulation of the knowledge needed to<br />

prevent rejection of the final system. To compare with another field, a participatory design approach has<br />

recently been used in individual education, particularly with an increasing interest in experiential- and<br />

learner-centered education; users take responsibility for their own learning. In other words, there has<br />

been some recognition that learners as well as teachers should participate in deciding what and how the<br />

teachers are expected to teach (Joyes & Chen 2007).<br />

Failure in information systems implementation may originate from different sources, such as development<br />

failures and information system usage problems. User participation, better modeling tools, and better<br />

management of the developing project are means that could be used to reduce these risks in information<br />

systems projects. When user participation was emphasized during the design process, this showed to be<br />

a factor for a successful design and resulted in positive user satisfaction (Nykänen & Karimaa 2006).<br />

Using a participatory design approach in health IT-based systems design and implementation allows for a<br />

group of people with a common interest to exchange ideas and experiences within the group. It seems<br />

that during their collaboration and exchange of experiences and skills, the staff members widen their<br />

knowledge, or, in other words, increase the human capital. The participatory design approach, through<br />

the involvement of end-users in the design and implementation processes, provides them the opportunity<br />

to learn how to work and collaborate with one another in a new and common subject. They learn how to<br />

use each other’s ideas and viewpoints to improve the quality of work. Moreover, by exchanging their<br />

experiences with each other, they obtain new experiences. In a broader context, they might employ this<br />

new experience and knowledge in their future work in the organization. It can be claimed that in this way<br />

the human capital is increased.<br />

375


Bahlol Rahimi<br />

From another perspective, despite the importance of information and communication technologies for<br />

knowledge sharing, it has been known for a long time that the creation and diffusion of knowledge within<br />

organizations relies on the development of social networks, a shared system of meanings, and cultivation<br />

of shared values and norms. As part of this emphasis on the social aspects of knowledge creation and<br />

sharing, considerable attention has recently been directed toward the role played by social capital<br />

(Vimarlund et al 2001; Spender 1996). In fact, users may not have the time to exchange information<br />

during the ordinary work life, yet this kind of involvement brings the users together and creates the<br />

possibility for building a network. From a social capital perspective, the individuals’ involvement in the<br />

development process can become more important because such a system allows on-the-job IT-training<br />

as well as the creation of a new social network, thereby incorporating new knowledge and skills or<br />

improving old ones, while working. This harbors both, present and future organizational benefits, from the<br />

experience and specialization of employees (Vimarlund & Timpka 1998).<br />

The benefits of social capital has, however, been identified and considered as “optimal” when it is created<br />

by individuals who learn to trust one another so that they are able to make credible commitments and rely<br />

on generalized forms of reciprocity, contributing to an increase in a society’s productive potential. In our<br />

case, social capital is obtained through mutual relationships that are used to accomplish different joint<br />

activities. Like financial capital, social capital is a resource that helps to sustain a community and<br />

encourages collaboration and cooperation between members of groups for their mutual benefits<br />

(Huysman & Wulf 2004). To summarize, the introduction of a new health information systems in which all<br />

users are involved and where networks are used for collaboration at different levels of the healthcare<br />

setting environment, are examples of such assets. Further, the networks are usually groups of people<br />

with common interests. We believe there is a particular need to involve users and other experts in the<br />

early concept phase, where important decisions are made. It can be concluded that using a participatory<br />

design approach to involve end-users in the design of their own system or program, benefit both the<br />

individual and the organization. It can be concluded that the application of a participatory design<br />

approach in health IT-based systems may be fruitful for the organization. The involvement of the endusers<br />

in the design and implementation process will allow for the collection of their ideas, thus benefitting<br />

the developing system as well as the development of the work routines. In addition, a feeling of<br />

ownership could be reached through their participation in the design and implementation of the system.<br />

7. Conclusion<br />

The study was related to the topic of factors that influence the implementation and use of integrated<br />

HISs. The reason to focus on this topic was that though there had been a strong increase in<br />

implementation and use of information systems in healthcare setting, issues such as adoption, HISs<br />

effects, and factors influencing the implementation and use of integrated HISs still need to receive more<br />

attention and publication of such studies contribute to the emergence of an evidence-based health<br />

informatics which can be used as a platform for decisions by policy makers, executives, and clinicians<br />

Based on the result, it can be concluded that there is an increasing need to share knowledge and to find<br />

methods for evaluating the impact of investments and formulating indicators for success. It is therefore<br />

interesting to develop or extend evaluation methods that can be applied to this area with a multi-actor<br />

perspective in order to understand the effects, consequences, and prerequisites for the successful<br />

implementation and use of IT in healthcare.<br />

It can also be concluded that HISs be introduced to fulfil a high number of organizational, individualbased,<br />

and socio-technical goals at different levels. It is therefore necessary to link the objectives that<br />

these systems are designed to achieve with organizations’ short-term, middle-term, and long-term<br />

strategic goals. Another conclusion is that implementers and vendors have to direct more attention to<br />

what has been published in the area to avoid more failures in the future. The other conclusion is that if we<br />

want more evidence-based practice, we need more practice-based evidence.<br />

Acknowledgements<br />

I wish to thank Toomas Timpka, Vivian Vimarlund and Anna Möberg for their valuable help to conduct<br />

this study.<br />

References<br />

Adi, A. (2000) Healthcare information systems: challenges of the new millennium. Hershey, PA, Idea Group<br />

Publishing, USA.<br />

376


Bahlol Rahimi<br />

Anderson, J. G. (1997) “Clearing the way for physicians? Use of clinical information systems”, Communications of<br />

the ACM. Vol 40. No. 8, pp 83-90.<br />

Blundell, R. Dearden, L. Meghir, C. and Sianesi, B. (1999) “Human capital investment: The returns from education<br />

and training to the individual, the firm and the economy”. Fiscal Studies Vol 20, No. 1, pp 1–23.<br />

Birkmann, C. Demski, H. and Engelbrecht, R. (2006) “Introducing patient cards in clinical routine: evaluation of two<br />

research projects”. Methods of Information in Medicine. Vol 45, No. 1, pp 73-8.<br />

Cohen, D. and Prusak, L. (2001) “In good company: How social capital makes organizations work.” Harvard<br />

Business School Press, Boston Massachusetts.<br />

Diamond, J. and Khemani, P. (2005) Introducing financial management information systems in developing countries.<br />

IMF Working Paper, Fiscal Affairs Department, WP/05/196.<br />

Faber, M.G.(2003) “Design and introduction of an electronic patient record: how to involve users?” Methods of<br />

Information in Medicine. Vol 42, No. 4, pp 371-5.<br />

Huysman, M. H. and Wulf, V. (2004) “Social capital and information technology”. MIT Press: Cambridge.<br />

Joyes, G. and Chen, Z. (2007) “Researching a participatory design for learning process in an intercultural context”,<br />

International Journal of Education and Development using Information and Communication Technology, Vol 3,<br />

No. 3, pp 78-88.<br />

Kaplan, B. (2000) “Culture counts: How institutional values affect computer use”, M. D. Computing. Vol 17, No.5, pp<br />

23–26.<br />

McKean, E. (2005) “Human capital” in The New Oxford American Dictionary, second edition. Oxford University<br />

Press. Oxford Reference Online.<br />

Nykänen, P. and Karimaa, E. (2006) “Success and failure factors in the regional health information system design<br />

process--results from a constructive evaluation study”, Methods of Information in Medicine. Vol 45, No. 1, pp<br />

85-9.<br />

Ostrom, E. (2000) “Social capital and development projects.” Progress in Planning Vol 54, No. 2, pp 73-132.<br />

Paré, G. and Elam, J. J. (1998) “Introducing Information Technology in the Clinical Setting: Lessons Learned in a<br />

Trauma Center”, International Journal of Technology Assessment in Health Care, Vol 14. No.2, pp 331-343.<br />

Pilemalm, S. and Timpka, T. (2007) “Third generation participatory design in health informatics-Making user<br />

participation applicable to large-scale information system projects”, Journal of biomedical informatics, Vol 41,<br />

No. 2, pp 327-39.<br />

Poon, E. G. Blumenthal, D. Jaggi, T. Honour, M. M. Bates D. W. and Kaushal, R. (2004) “Overcoming barriers to<br />

adopting and implementing computerized physician order entry systems in U.S. hospitals”, Health Affaires<br />

(Millwood). Vol 23, No.4 , pp 184-90.<br />

Preece, J. (2002) “Supporting community and building social capital.” Communications of the ACM. Vol 45, No. 4, pp<br />

37-39<br />

Putman, R. D. (1993a) “The Prosperous Community: Social Capital and Economic Growth.” American Prospect, Vol<br />

13, No. 1, pp 35-42.<br />

Putman, R. D.(1993b) “Making democracy work: Civic traditions in modern Italy”, Princeton University Press,<br />

Princeton, NJ.<br />

Rahimi, B. Moberg, A. Timpka, T. and Vimarlund, V. (2008) “Post-implementation survey on using of integrated<br />

computerized patient record system”, MIE 2008 conference Proceeding, Gothenburg, Sweden, May.<br />

Rahimi, B. and Vimarlund, V. (2007) “Methods to evaluate health information systems in healthcare settings: A<br />

literature review”, Journal of Medical Systems. Vol 31, No. 5, pp 397–432.<br />

Schuler, D. and Namioka, A. (1993) “Participatory design: Principles and practices. Lawrence Erlbaum Associates,<br />

Hillsdale, NJ.<br />

Sengstack, P. P. and Gugerty, B. (2004) “CPOE systems: success factors and implementation issues”. Journal of<br />

Healthcare Information Management, Vol 18, No. 1 pp 36-45.<br />

Sjöberg, C. and Timpka, T. (1998) “Participatory design of information systems in health care”, Journal of the<br />

American Medical Informatics Association, Vol 5, No. 2, pp 177-83.<br />

Smullen. J. and Hand. N. (2005) “Human capital” in A Dictionary of Finance and Banking. Oxford University Press.<br />

Oxford Reference Online.<br />

Spender, J.C. (1996) “Organizational knowledge, learning and memory: three concepts in search of a theory.”<br />

Journal of Organizational Change Management, Vol 9, No. 1, pp 63-78.<br />

Van der Meijden, M. J. Tange, H. Troost, J. and Hasman, A. (2001) “Development and implementation of an EPR:<br />

how to encourage the user?” International Journal of Medical Informatics. Vol 64, No. 2-3, pp 173-85.<br />

Van Der Meijden, M. J. Tange, H. J. Troost, J. and Hasman, A. (2003) “Determinants of success of inpatient clinical<br />

information systems: A literature review”. Journal of the American Medical Informatics Association. Vol 10, No.<br />

3, pp 235–243.<br />

Vimarlund, V. Eriksson, H. and Timpka, T. (2001) “Economic motives to use a participatory design approach in the<br />

development of public-health information systems.” Medinfo. Vol 84, pp 768-72.<br />

Vimarlund, V. and Timpka, T. (1998) “Participatory design in economic terms: an economic discussion.” PDC 98<br />

proceeding of Participatory design conference. Washington USA. pp 11-17.<br />

Weng, C. McDonald, D. W. Sparks, D. McCoy, J. and Gennari, J. H. (2007) “Participatory design of a collaborative<br />

clinical trial protocol writing system.” International Journal of Medical Informatics. Vol 76, Suppl 1, pp 245-51.<br />

Wild, E. L. Hastings, T. M. Gubernick, R. Ross, D.A. and Fehrenbach, S. N. (2004) “Key elements for successful<br />

integrated health information systems: lessons from the States”. Journal of Public Health Management Practice,<br />

November (Suppl), pp 36-47.<br />

377


Relationship Between Organizational Factors and RFID<br />

Adoption in Iranian Gas Company Central Warehouses<br />

Farnaz Rahimi 1, 2 and Gholamabas Arabshahi 2<br />

1 Alzahra University, Tehran, Iran<br />

2 Gas company, Mashhad, Iran<br />

farnaz12200@yahoo.com<br />

a_arabshahi@ent.ut.ac.ir<br />

Abstract: This paper examined organizational factors that influence adoption of RFID (Radio Frequency<br />

Identification) technology in IRAN Gas company central warehouses. RFID has the distinct advantage of collecting<br />

data at some distance from the actual product, with no direct line of sight and without labor. Such a data automation<br />

system for real-time tracking, safety monitoring and overall warehouse operation leads to real-time visibility and<br />

tracking of assets and inventory that is necessary for a warehouse. In this paper we examine organizational factors<br />

(Structural: Size and resources and Centralization and Cultural: Innovativeness) associated with the adoption of<br />

RFID in gas company warehouses. Using a survey questionnaire we collected data from all 30 Gas company central<br />

warehouses employees(our unit of analyses) in IRAN .We found that organizational size do not show any<br />

relationship With RFID adoption and Centralization finantional recourses and Innovativeness are positively and<br />

significantly associated with the adoption of RFID . The implications of these and other findings are discussed.<br />

Keywords: organizational factors, warehouse, RFID<br />

1. Introduction<br />

Using radio waves, RFID technology can automatically identify objects. With RFID tags, objects can be<br />

tracked automatically by radio readers providing greater inventory visibility and improved business and<br />

control processes (Ngai et al.2007 Nystrom et al.2006) RFID’s main ability is to capture more data<br />

automatically without human intervention, in almost real time, so it can provide a more dynamic control<br />

environment for an organization.<br />

While RFID has been discussed in the literature as a technology that can provide several advantages:<br />

enhanced warehouse operations efficiency , better of inventory control and monitoring (McFarlane and<br />

sheffi 2003, Higgins,2006) , reduced labor costs (Kinsella,2003) , inventory obsolescence material<br />

handling costs reduction (McFarlane and sheffi 2003), quality control (Hoffman,2006), reductions in outof-stock<br />

and delivery safety stock (Kinsella,2003)uncertainty of product availability reduction (Asif and<br />

Mandviwalla 2005), to its adopters ,the RFID adoption rate is not growing as fast as expected (Nystrom et<br />

al.2006. Boucher 2007).This suggests more effort is necessary to understand the process of adoption of<br />

the technology and to identify factors affecting the RFID adoption decision (Ngai et al2008).<br />

According to Rogers et al. (1996) companies that used more warehousing information technologies had<br />

better performance in the areas of quality and productivity improvements and cycle time reductions than<br />

the low user group. Past researches show that, organizational factors often have significant influence on<br />

the process of technology adoption.(Ko et al.2000; Lefebvre et al. ,1991 ; Masters et al. ,1992<br />

;O'Callaghan et al. ,1992 ;Sullivan ,1990; Premkumar et al .1997).<br />

In this paper we study the relationships between organizational factors (size, financial resources and<br />

centralization) as structural characteristic and (organizational culture) and RFID adoption in the Gas<br />

company warehouses.<br />

The rest of the paper is organized as follows: Section 2 introduces RFID, and the literature review of<br />

RFID adoption studies .In Section3, we present the research model and hypotheses. This is followed by<br />

the description of the research methods used in data collection. Section 5 the findings will be presented.<br />

2. Literature review<br />

2.1 RFID<br />

RFID is a common term for technologies and systems that use radio waves to automatically identify<br />

people or objects.<br />

378


Farnaz Rahimi and Gholamabas Arabshahi<br />

RFID typically consists of three basic components: tags (transponders), readers, and middleware (Asif et<br />

al.2005 ; Preradovic.2007).<br />

A tag usually containing a microchip and an antenna that are attached to or embedded in an object<br />

(Curtin 2007) The microchip contains identification information and may have other application data (e.g.<br />

price, cost, location, and manufacture date, etc.)(Chao 2005)<br />

A reader (also known as an interrogator) sends out a radio signal and abets the tag to broadcast the data<br />

contained on its chip. The reader then change the radio waves returned from the tag into digital data and<br />

forward them to a computer system (High jump Software, 2004; Zebra Tech- neologies, 2004).<br />

2.2 Reviews of related studies<br />

Earlier studies on RFID adoption have focused on a variety of factors associated in adoption process:<br />

Hoske (2004) study the cost factor .Jones, Clarke-Hill, Shears, Hillier, and Comfort (2004b) looked at<br />

privacy and public policy with regards to RFID.<br />

Adam (2004) investigated benefits, costs, standards, and environment; four main factors associated<br />

RFID adoption.<br />

Jones M .et al. (2005) identified benefits, costs, standards, privacy, and the power of retailers as factors.<br />

In their study to identify the factors may influence RFID adoption in South African retail organizations,<br />

Brown and Russell (2007) found that the RFID adoption intention was explained by technological factors<br />

(i.e., relative advantage ,compatibility complexity ,and cost) ,organizational factors (i.e., top management<br />

attitude ,information technology expertise, organization size organizational readiness), and external<br />

factors(i.e., competitive pressure, external support, and existence of change agents.)<br />

Reviewed prior studies, Schmitt et al. (2007) categorized 25 adoption factors in the technology,<br />

organizational, and environment groups. They concluded five most important factors affecting RFID<br />

adoption and diffusion in the automotive industry: compatibility, costs, complexity, performance, and top<br />

management support. Most of these factors belong to the technology characteristics group.<br />

Based on previous literature, there are variety of factors related to technology adoption that we can<br />

categorize them into four groups: environmental, technological, inter-organizational and organizational.<br />

Each group contains different variables in different studies.<br />

In this research we analyze organizational factors comprising four variables discussed below:<br />

3. Research model and hypotheses<br />

3.1 Organizational size<br />

Organizational size has an important impact on the adoption of technological adoption (Kimberly and<br />

Evanisko, 1981; Fama and Jensen, 1983; Kennedy, 1983; Damanpour,1992)<br />

In some studies a positive relationship between firm size and adoption has been concluded (Tornatzky,<br />

1990;Grover, 1993;Premkumar1999; Damanpour, 1992; Moon and Bretschneider 1997).Because of the<br />

cost of RFID tags and other related equipments ,only large companies have the financial resources to<br />

invest in RFID(Wang et al,.2006.) In addition, greater formalization (Pugh et al 1969), greater task<br />

specialization (Blau1970), and more complex forms of communications (Haveman 1993) are<br />

characteristics; help larger organization adopt new technologies.<br />

Although counterarguments exist that smaller organizations have more flexibility due to fewer levels of<br />

bureaucracy, thus increasing their ability to adopt technology more quickly (Patterson et al 2003), few<br />

empirical results have shown that smaller firms tend to be more innovative (Salavou, et al. 2004)<br />

In this study, organizational size is defined as work force size (number of employees) (Kimberly and<br />

Evanisko 1981;Teo and Tan ,1998) and the number of items categorized in each warehouse.<br />

We expect that organization size will have a positive influence on RFID adoption:<br />

379


Farnaz Rahimi and Gholamabas Arabshahi<br />

H1: There is a positive relationship between size of organization in terms of number of employees and<br />

number of items in each warehouse and RFID adoption.<br />

3.2 Financial resources<br />

Financial resources of an organization are another predictors of innovation (Hage and Aiken 1967; Hurley<br />

and Hult 1998).<br />

Financial resources facilitate the adoption of technology and provide organizations with the capacity to<br />

accept risk inherent in investing in new technology (Grover and Goslar 1993).<br />

The measure we used in this study is the budget allocated for new technology acquisition.<br />

H2: There is a positive relationship between the amount of financial resources and RFID adoption.<br />

3.3 Centralization<br />

Centralization has been defined as the degree that power is distributed among all the positions within an<br />

organization (Hage and Aiken 1967; Schminke et al.2000).<br />

It consists of two dimensions :(Hage and Aiken 1967; Schminke, et al.2000)<br />

1)Participation in decision making which shows how much members of the organization, participate in<br />

organizational decisions(Hage and Aiken 1967), and 2) Authority hierarchy which refers to" the amount of<br />

autonomy employees have in relation to deferring to superiors when making decisions about day to day<br />

tasks" (Tata and Prasad 2004, p. 251)<br />

Rogers (1995) suggested that centralization is negatively associated with organizational adoption. This is<br />

empirically supported by prior studies (e.g Hage, 1969; Moch and Morse, 1977; Kimberly and Evanisko,<br />

1981; Pierce and Delbecq, 1977;Grover 1993; Grover and Goslar 1993)<br />

A reason for this relationship is that in a centralized structure, top management is not aware of<br />

operational- level problems and con not suggests relevant innovations to meet their needs. Another<br />

reason is that the more power is concentrated at the top by a few strong leaders, the more new ideas and<br />

innovations are restricted .( Yoon,2009)<br />

However, some researchers found a positive relationship between centralization and technology adoption<br />

.(Ettlie et al. 1984).They found that centralized decision making would facilitate faster and more efficient<br />

adoption because lower level management resistance would not impede top management support.<br />

Evidence has also supported that decentralized organizations are more likely to adopt technological<br />

innovations.<br />

In this study we ask the respondents to state how much they are interfered to every day decision making<br />

and how much their boss intervene to their tasks.<br />

Based on these arguments and previous findings, the following hypothesis is proposed in this research:<br />

H3: There is a negative relationship between the degree of centralization of the organization and<br />

adoption of RFID.<br />

3.4 Organizational culture<br />

There are over 54 definitions of organizational culture in literature. Verbeke et a l.(1998). “A system of<br />

shared norms and behaviors that are learned by members of the organization and shape their way of<br />

doing” (p. 313). Sutherland and Morieux (1988) defined organizational culture as “a set of emotional<br />

conditions, views of the past and standards of behavior belonging collectively to an organization (p. 43).<br />

Organizational culture can explain why some organizations adopt new technology quicker than their<br />

competitors. (Glisson and James 2002).<br />

380


Farnaz Rahimi and Gholamabas Arabshahi<br />

We ask the respondents about management’s attitude towards innovations and new ideas and also<br />

respondent willingness to search and seek innovations. (Hurley and Hult1998)<br />

H4: There is a positive relationship between innovative organizational culture and adoption of RFID.<br />

4. Research design<br />

The propositions in last Section formed the basis for carrying out the investigation. This study is an<br />

applicable-descriptive research. We use questionnaire which developed by adopting measures from<br />

several sources (Hurley and Hult 1998;Hage and Aiken 1967; Jaworski and Kohli1993; Matsuno et al.<br />

2002; Kimberly and Evanisko1981).<br />

A5-point Likert scale was used for each item, anchored by Strongly Agree at one end to Strongly<br />

Disagree at the other. The target population for this study was Gas company central warehouses<br />

employees in all the states in IRAN. Due to resource constraints, we chose a sample, using random<br />

sampling and mail the questioners to 145 respondents.120 questioner were returned and used for our<br />

analysis.<br />

4.1 Reliability<br />

Cronbach’s Coefficient Alpha in a small sample size was 0/93 which means the questioner has a high<br />

degree of reliability.<br />

5. Data analysis<br />

Using SPSS software, we specify Correlation Coefficient between RFID Adoption and other factors<br />

(organizational size, centralization, financial resources, and innovative culture) as follows:<br />

Table 1: Results of analyzed data<br />

RFID Adoptionorganizational<br />

size<br />

RFID Adoption -<br />

centralization<br />

RFID Adoption –<br />

financial resources<br />

RFID Adoption –<br />

innovative culture<br />

sample<br />

size<br />

120<br />

120<br />

120<br />

120<br />

Correlation<br />

Coefficient<br />

-0.069<br />

0.461<br />

0.208<br />

0.599<br />

Statistic<br />

0.75<br />

5.64<br />

2.31<br />

8.13<br />

6. Findings<br />

As mentioned in the table, the results are as followed:<br />

P-value<br />

0.452<br />

0/000<br />

0.023<br />

0.000<br />

α<br />

0.05<br />

0.05<br />

0.05<br />

0.05<br />

result<br />

there is no<br />

relation between<br />

RFID Adoptionorganizational<br />

size<br />

there is a positive<br />

relationship<br />

between RFID<br />

Adoptionorganizational<br />

size<br />

there is a positive<br />

relationship<br />

between RFID<br />

Adoptionorganizational<br />

size<br />

there is a positive<br />

relationship<br />

between RFID<br />

Adoptionorganizational<br />

size<br />

Organizational size does not have impact on RFID adoption. All other structural variables (finantional<br />

resources, centralization) and also innovative culture of the organization have a positive relationship with<br />

RFID adoption. So the propositions number 2 and 4 (financial resources and innovative culture) are<br />

proved and the other 2 hypothesis (size and centralization) rejected.<br />

These findings can guide the decision makers. It shows that consideration of innovativeness and<br />

allocating financial recourses to new technology can speed up the rate of adoption. Because Gas<br />

381


Farnaz Rahimi and Gholamabas Arabshahi<br />

Company is governmental in Iran, centralization is unavoidable. In such situations, decision makers<br />

should concern financial resources and innovative culture more.<br />

Acknowledgements<br />

We highly appreciate Khorasan Razavi Gas Co for their technical and financial support of this research.<br />

References<br />

Adam ,K.(2004). The South African retailers’ propensity to adopt radio frequency identification technology.<br />

Unpublished Honours. Technical Report, University of Cape Town<br />

Asif Z.,M .Mandviwalla, Integrating the supply chain with RFID: a technical and business analysis, Communication of<br />

AIS1 5(24), 2005,pp.393–426.<br />

Blau, P.M., A formal theory of differentiation in organizations. American Sociological Review 35,1970,pp.201–218.<br />

Boucher, R.,RFID adoption is lagging,eWeek.24(1)(2007)25.<br />

Brown, I. Russell, J. ,Radio frequency identification technology :an exploratory study on adoption in the South African<br />

retail sector ,Int. J. Inform .Manage.27.(4) (2007) 250-265.<br />

Chao ,C.C. Yang,J.M. Jen,W.Y. ,Determining technology trends and forecasts of RFID by a historical review and<br />

bibliometric analysis from1991to2005.Technovation27(5)(2007)268–279.<br />

Curtin ,J. Kauffman, R.J. Riggins, F.J., Making the ‘MOST’ out of RFID technology :a research agenda for the study<br />

of the adoption, usage and impact of RFID Inform.Technol.Manage.8(2)(2007)87–110<br />

Damanpour, F.,1992. Organizational size and innovation. Organization Studies 13,375 – 402.<br />

Ettlie, John E., William P. Bridges and Robert D. O’Keefe (1984), “Organizational Strategy and Structural Differences<br />

for Radical versus Incremental Innovation Management Science, 30 (6), 682-95.<br />

Fama,E. ,Jensen, M.,1983. Separation of ownership and control. Journal of Law and Economics 26,301–325.<br />

Glisson, Charles and Lawrence R. James (2002), “The Cross-level Effects of Culture and Climate in Human Service<br />

Teams,” Journal of Organizational Behavior, 23,767-94.<br />

Grover, V. (1993) An empirically derived model for the adoption of customer-based interorganizational systems,<br />

decision science , 24, 3, 603-640.<br />

Grover, V., and Goslar, M.D. (1993) The initiation, adoption, and implementation of telecommunications technologies<br />

in U.S. organizations, journal of management information systems, 10, 1, 141.<br />

Hage, Jerald and Michael Aiken (1967), “Relationship of Centralization to Other Structural Properties,” Administrative<br />

Science Quarterly, 12, 72-92.<br />

Hage, J. (1969) Routine technology, social structure, and organizational goals ,administrative science quarterly, 14,<br />

June, 1969, 368-379.<br />

Haveman ,H. A., Organizational size and change :diversification in the savings and loan industry after deregulation,<br />

Administrative Science Quarterly 38(1),1993,pp.20–50.<br />

Higgins.L.N. Cairney,T., RFID opportunities and risks ,Journal of Corporate Accounting &Finance 17(5),2006,pp.51–<br />

57.<br />

High jump Software,2004.The true cost of radio frequency identification (RFID).High jump Software Free<br />

Report.URL: <br />

Hoffman,W. Metro moves onRFID,TrafficWorld270(4),2006,p.18.<br />

Hoske, M.(2004).RFID: Adoption increases despite costs. Control Engineering, 51(7),46–47.<br />

Hurley, Robert F. and G. Thomas Hult (1998), “Innovation, Market Orientation, and Organizational Learning: An<br />

Integration and Empirical Examination,” Journal of Marketing, 62 (July), 42-54<br />

Jaworski, Bernard J. and Ajay K. Kohli (1993), “Market Orientation: Antecedents and Consequences,” Journal of<br />

Marketing, 57 (July), 53-70.<br />

Jones ,M. ,Wyld,D. ,&Totten,J. (2005).The adoption of RFID technology in the retail supply chain. The Costal<br />

Business Journal, 4(1)29-42.<br />

Jones,P.,Clarke-Hill, C.,Shears,P ,Hillier,D .,&Comfort,D .(2004). Radio frequency identification in retailing and<br />

privacy and public Policy issues.Management Research News, 27(8/9),46–56.<br />

Kennedy,A.M. ,1983. The adoption and diffusion of new industrial products: a literature review .<strong>European</strong> Journal of<br />

Marketing17,31–88.<br />

Kimberly, John R. and Evanisko , Michael J. (1981), “Organizational Innovation: The Influence of Individual,<br />

Organizational, and Contextual Factors on Hospital Adoption of Technological and Administrative Innovations,”<br />

Academy of Management Journal, 24 (4), 689-713.<br />

Kinsella.B.,The Wal-Mart factor,Industrial Engineer35(11),2003 ,pp. 32–36.<br />

Ko,E. Kincade, DH ,Brown JR. Impact of business strategy on the adoption of quick response technologies :the<br />

apparel industry experience .IntJOper ProdManage2000;20(9):1093-111.<br />

Lefebvre LA, Harvey J, Lefebvre E. Technological experience and technology adoption decisions in small<br />

manufacturing firms .R&D Manag 1991;21 (3):241-9.<br />

Masters JM ,Allenby GM, Lalonde BJ, Maltz A .On the adoption of DRP .JBus Logist1992;13(1):47-67.<br />

Matsuno, Ken, John T. Mentzer and Aysegul Ozsomer (2002), “The Effects of Entrepreneurial Proclivity and Market<br />

Orientation on Business Performance Journal of Marketing, 66 (3), 18-32<br />

Mc Farlane ,D.Sheffi,Y.,The impact of automatic identification on supply operations chain,International Journal of<br />

Logistics Management14(1),2003,pp.1–17.<br />

Moch, M., and Morse, E. (1977) Size, centralization and organizational adoption of innovations,Amer.social.Rev, 92,<br />

Oct, 1977, 716-725.<br />

382


Farnaz Rahimi and Gholamabas Arabshahi<br />

Moon,M.J.,Bretschneider,S.,Can state government actions affect innovation and its diffusion?:an extended<br />

communication model and empirical test Technol.Forecast.Soc.Change54(1)(1997)57–77.<br />

Ngai.E.W.T. ChengT.C.E. Au,S. Lai,K. ,Mobile commerce integrated with RFID technology in a contained<br />

depot,Decis.SupportSyst.43(1)(2007)62–76.<br />

Ngai,E.W.T. Moon,K.K.L. Riggins,F.Yi.J.,C.Y.,RFID research: an academic literature review(1995–2005) and future<br />

research directions, Int .J.Prod.Econ.112.520-510(2008)(2).<br />

Nystrom,N.C.Lin,H.A.Yu,H.C. Wu,M.A,Challenges to global RFID adoption, Technovation 26 (12) (2006)1317–1323.<br />

O'Callaghan R ,Kaufmann PJ, Konsynski BR. Adoption correlate and share effects of electronic data interchange<br />

systems in marketing channels. JMark 1992;56(2):45-56.<br />

Patterson, Kirk A., Curtis M. Grimm, and Thomas M. Corsi (2003), “Adopting New Technologies for Supply Chain<br />

Management,” Transportation Research Part E 39 (march),95-121.<br />

Pierce, J.L., and Delbecq, A.L. (1977) Organizational structure, individual attributes and innovation, journal of<br />

management information system, 2, 1, January, 27-37.<br />

Premkumar,G.,Ramamurthy,K.,Crum,M.,1997.DeterminantsofEDI adoption in the transportation industry .<strong>European</strong><br />

Journal of InformationSystems6,107–121.<br />

Premkumar,G.,Roberts,M.,Adoption of new information technologies in rural small<br />

businesses,Omega27(4)(1999)467–484.<br />

Preradovic ,S. ,N.C.Karmakar ,Modern RFID readers ,Microwave Journal 50(9)(2007).<br />

Pugh,D.S. Hickson,D.J.Hinings,C.R.Turner,C. ,The context of organizational structure,Administrative Science<br />

Quarterly 14 1969,pp.91–114.<br />

Rogers,dale.daugherty.s.pj.Elllinger.a.e."the relationship between information technology and warehousing<br />

performance".logestics and transportation review.vol 32.Desember 1996.<br />

Rogers, E. (1995),diffusion of innovations, (4th ed.) The Free Press, New York.<br />

Salavou, H., G. Baltas and S. Lioukas (2004), “Organisational Innovation in SMEs <strong>European</strong> Journal of Marketing, 38<br />

(9/10), 1091-112.<br />

Schmitt,P., Thiesse ,F.,E.Fleisch ,Adoption and diffusion of RFID technology in the automotive industry ,in: H.Osterle,<br />

J.Schelp ,R. Winter(Eds.),Proceedings of the15th <strong>European</strong> <strong>Conference</strong>on Information Systems, St<br />

Gallen,Switzerland,2007.<br />

Schminke, Marshall, Maureen L. Ambrose and Russell S. Cropanzano (2000), “The Effect of Organizational<br />

Structure on Perceptions of Procedural Fairness Journal of Applied Psychology, 85 (2), 294-304<br />

Sullivan PC. A study of the adoption of quick response in the United States apparel manufacturing industry<br />

.Unpublished doctoral dissertation .New YorkUniversity;1990.<br />

Sutherland, Ewan and Yves Morieux (1988), “Effectiveness and Competition – Linking Business Strategy,<br />

Organizational Culture and the Use of Information Technology,” JIT, 3 (1), 43-47<br />

Tata, Jasmine and Sameer Prasad (2004), “Team Self-Management, Organizational Structure, and Judgments of<br />

Team Effectiveness,” Journal of Managerial Issues16 (2),251.<br />

Tan, M., and Teo, T.S.H. (1998) Factors influencing the adoption of the Internet,international journal of electronic<br />

commerce, 2, 3, 5-18.<br />

Tornatzky .M.L.G., The processes of technological innovation, Lexington Books,Fleischer,1990.<br />

Verbeke, Willem, Marco Volgering, and Marco Hessels (1998), “Exploring the Conceptual Expansion within the Field<br />

of Organizational Behaviour Organizational Climate and Organizational Culture,” Journal of Management<br />

Studies, 35 (3), 303-29<br />

Wang,W., Yuan, Y., Wang, X., Archer, N.,RFID implementation issues in China: Shanghai port cast study ,J.<br />

InternetComm. 5(4) (2006) 89–103.<br />

Yoon,Tom.(2009).an empirical investigation of factors affecting organizational adoption of virtual world. ProQuest.<br />

Zebra Technologies, 2004 .RFID :the next generation of AIDC . Zebra TechnologiesWhitePaper.URL:<br />

<br />

383


Business Intelligence as Decision Support in Business<br />

Processes: An Empirical Investigation<br />

Ari Riabacke 1 , Aron Larsson 1,2 and Mats Danielson 1<br />

1<br />

Stockholm University, Kista, Sweden<br />

2<br />

Mid Sweden University, Sundsvall, Sweden<br />

ariria@dsv.su.se<br />

aron.larsson@miun.se<br />

mad@dsv.su.se<br />

Abstract: Our concern in this paper is the role of business intelligence systems and the perceived business value of<br />

implemented systems and their contribution to facilitate the fulfilment of organizational goals. The study builds upon a<br />

survey answered by 43 respondents from different large companies in Scandinavia. The survey used questions on<br />

how visions, objectives, strategies are supported by BI systems, on how business values are derived from such<br />

systems, and on how design and implementation issues affect the solutions. The overall conclusion of the study is<br />

that there are markedly different levels of problems in the areas, most problems being found in integration of BI<br />

information and decision processes, and that there is room for large improvements and further work within everything<br />

from implementation to requirements engineering for business intelligence decision support systems.<br />

Keywords: business intelligence, decision support, business value, empirical study<br />

1. Introduction<br />

During the latter decades organizational decision making has become increasingly complex due to the<br />

amount of accessible information that managers wish to incorporate in decision making activities has<br />

increased dramatically. Given the integration of information systems into business processes,<br />

organizational decision making may often experience a substantial amount of environmental pressure<br />

due to, for instance, information overload and sharper time constraints (Rasmussen 2000, Speier and<br />

Morris 2000). At present, one popular approach to deal with the increasing amount of data in<br />

organizations is to build data warehouses and BI-systems.<br />

Systems or methods for supporting decisions come in many forms. For example, decision analytical<br />

systems, focusing on supporting problem structuring and evaluations of alternative courses of action,<br />

optimization tools focusing on finding an optimal solution for a well-defined system of constrained<br />

variables, or business intelligence systems (BI) aiding in the gathering and analysis of business<br />

information. See (Turban et al. 2010) for a comprehensive treatment of business intelligence. In this<br />

respect, business intelligence systems may be viewed as information systems with special focus on<br />

providing accessible business data, i.e., they can be viewed as type of decision support system with the<br />

capability of (easily and quickly) providing reliable and up-to-date information or key figures about the<br />

organization.<br />

Since the 1990s, such BI systems typically employ three different technologies for supporting decision<br />

making in an organization: data warehouses (DW) for the gathering of business data, data mining (DM),<br />

and on-line analytical processing (OLAP) for data analysis (Shim et al. 2002). The benefits of a decision<br />

support system may further vary. Already in 1981, Keen summarized the potential benefits of a decision<br />

support system (DSS) as they enable the following: facilitate a more rational treatment of decisions<br />

problems; yield a better understanding of the business; yield faster response to unexpected situations;<br />

facilitate ad-hoc analyses; yield new insights and learning; improve communication; enhance control;<br />

enhance team work; reduce time spent on planning; and better use of resources (Keen 1981). This<br />

basically remains true even today. However, the decision support function enabling better decision<br />

making and other more advanced uses of a BI system such as facilitating business strategies is hard to<br />

measure (Watson and Wixom 2007). The results of a previous study presented in (Stenfors et al. 2007),<br />

according to managers the strongest incentives for investing in a DSS are to increase the effectiveness of<br />

operational processes and to support strategic work and decision making. These benefits are indeed of a<br />

more qualitative kind, and the decision to invest in a DSS is therefore often based rather on estimated or<br />

perceived values than on actual costs. Further, the results from the abovementioned study indicate that<br />

managers view DSS and BI systems as support instruments to business process management. However<br />

enabling this support does call for the BI system to be systematically aligned or integrated with business<br />

processes to a greater extent (Bogdana et al. 2009). The more recent and wider concept of BPM<br />

(business performance management) is related to business intelligence as the utilization of BI systems<br />

384


Ari Riabacke et al.<br />

can be viewed as an element of BPM, but where the use is supposed to be aligned with organizational<br />

strategy and determined by a program for how to implement this strategy (Frolich and Ariyachandra<br />

2006).<br />

But how does this need for a DSS manifest itself? Recent years have witnessed a remarkable increase of<br />

companies investing in BI systems, see (Hannula and Pirttimäki 2003). This makes it interesting to study<br />

the use and knowledge of the effects on the businesses of the companies utilizing, or having the ability to<br />

utilize, such systems. Further, studies of BI systems as a support for organizational decision making are<br />

relatively scarce. With respect to information systems in general, several studies have been carried out in<br />

order to identify factors behind a widespread use and end-user acceptance of a system within an<br />

organization. These factors are often user-related, such as user attitudes, their previous experiences, and<br />

user participation in development, but also context-related, such as top management support and<br />

conditions facilitating the use of information systems; see (Bajaj and Nidumolu 1998, Sabherwal et al.<br />

2006, Riabacke 2006).<br />

It is well-known that up to 50%, some say even 70%, of all BI implementations fail in one or more<br />

respects, not delivering the value-for-investment expected at the outset. While there are studies on the<br />

pitfalls of developing and realizing BI systems, this study focuses on systems already in operation, their<br />

organisational context, and hurdles on the path to efficient and effective use of BI information. We will, in<br />

the paper conform to the common use of the concept of system, denoting the technical part of a BI effort.<br />

From a systems science perspective, the system boundaries are indeed much wider, encompassing the<br />

organisation and its control mechanisms. Not to interfere with colloquial interpretations of the system<br />

concept, we will refer to this wider concept as the BI scope.<br />

BI is not about technology, it is much more. It is about providing an organisation with support for better<br />

decision making. Our hypothesis, from literature and interviews, is that far too much focus in the<br />

development and implementation of BI solutions in organizations is put on the technical parts of the<br />

system. Much of the effort is thus on the right hand side of the model shown in Figure 1, i.e. on the<br />

design and building of data warehouses, and on creating different types of reports. In comparison less<br />

effort seems to be put on structuring and understanding business processes and gaining insights into the<br />

decision making activities involved in these processes that are to benefit from the BI systems. In other<br />

words, the level of integration of BI systems into business processes could be much higher, yielding a<br />

better business value of the investments made in the system.<br />

Figure 1: Objects of interaction in the use of BI applications and systems<br />

Thus, the main area of concern in this paper is to what extent this view is supported by studying the<br />

perceived values of BI systems use and the views on system design and implementation.<br />

2. Study outline and method<br />

The study was conducted using a quantitative method based upon a questionnaire. Of 1085 persons<br />

considered for the survey and accessible through a large and collaborative consultancy firm, 105 were<br />

selected for participation. Of these 105, 43 completed the questionnaire, yielding an effective completion<br />

response rate of 41%. Thus, 43 respondents from different large Scandinavian companies employing BI<br />

systems for decision making support, having a leading role in these, fully answered the questionnaire in<br />

which they had to estimate how well different statements regarding their BI solutions corresponded to<br />

385


Ari Riabacke et al.<br />

their own perception of the reality in their own business. The respondents came from various business<br />

segments according to Table 1.<br />

Table 1: Respondents’ business segments<br />

Business segments Count<br />

IT 5<br />

Logistics 4<br />

Banking and insurance 15<br />

Real estate and construction 1<br />

Consumer products 2<br />

Energy 3<br />

Pharma and health care 1<br />

Industry 6<br />

Public sector 4<br />

Other 2<br />

TOTAL 43<br />

Table 2: Respondent’s role (some have more than one role)<br />

Role Count<br />

CEO 3<br />

CIO 3<br />

CFO 6<br />

Other management 3<br />

Head of Business Segment 4<br />

IT Management 7<br />

System owner 1<br />

Requirements officer 3<br />

System architect 4<br />

Developer 4<br />

Other 10<br />

Total 48<br />

The questions were divided into six categories 1 to 6, each comprising of six questions. The categories<br />

are:<br />

1: Visions, objectives, and strategies<br />

2: Business values from BI systems<br />

3: Requirements analysis and needs<br />

4: Change management<br />

5: Technical solutions<br />

6: Decision making support<br />

Thus, each category consists of 6 statements which the respondents were asked to assess. The<br />

respondents indicated how much they agreed with each statement on a 1-to-7 Likert scale, where 1 was<br />

strongly disagreeing and 7 was strongly agreeing. The statements were designed so that a higher value<br />

indicated better integration of BI into the business processes. One example of a statement in category 1<br />

is “BI has a role in our strategic processes”. See the appendix for the complete set of statements. For<br />

each category, the aggregated results (modal scores) of all respondents are shown in two spider<br />

diagrams (one for each category cluster). Thus, the area in the diagram represents the degree of BI<br />

integration in the category according to the responses.<br />

386


3. Results<br />

Ari Riabacke et al.<br />

The results are presented in two formats. The first format compares, for each category and question, the<br />

number of respondents who assessed a score of below 4 (the neutral score) to the number of<br />

respondents who assessed a score of above 4. These results are presented in three bar chart diagrams,<br />

one for each pair of categories. The other format is in the form of spider diagrams showing the modal<br />

score of each question, i.e. the most common assessment for each question. As there were more than<br />

one mode for some questions, the darker areas represent the result obtained from minimum modal<br />

scores and the brighter areas the maximum.<br />

3.1 Below neutral score vs. above neutral score<br />

2<br />

2<br />

1<br />

1<br />

5<br />

0<br />

Figure 2: Number of responds below (red) and above (green) the neutral score of 4 for categories 1-2<br />

30<br />

25<br />

20<br />

15<br />

10<br />

5<br />

0<br />

3.1 3.2 3.3 3.4 3.5 3.6 4.1 4.2 4.3 4.4 4.5 4.6<br />

Below 4<br />

Above 4<br />

Figure 3: Number of responds below (red) and above (green) the neutral score of 4 for categories 3-4<br />

35<br />

30<br />

25<br />

20<br />

15<br />

10<br />

5<br />

0<br />

1. 1. 1. 1. 1. 1. 2. 2. 2. 2. 2. 2.<br />

5.1 5.2 5.3 5.4 5.5 5.6 6.1 6.2 6.3 6.4 6.5 6.6<br />

Below 4<br />

Above 4<br />

Figure 4: Number of responds below (red) and above (green) the neutral score of 4 for categories 5-6<br />

387<br />

Below<br />

Above


3.2 Modal scores<br />

Ari Riabacke et al.<br />

1.1<br />

7<br />

6<br />

5<br />

4<br />

3<br />

2<br />

1<br />

0<br />

Figure 5: Spider diagram of modal scores for categories 1-2<br />

2.4<br />

2.5<br />

2.3<br />

2.6<br />

2.2<br />

2.1<br />

3.1<br />

7<br />

6<br />

5<br />

4<br />

3<br />

2<br />

1<br />

0<br />

Figure 6: Spider diagram of modal scores for categories 3-4<br />

4.4<br />

4.5<br />

4.3<br />

4.6<br />

4.2<br />

4.1<br />

5.1<br />

7<br />

6<br />

5<br />

4<br />

3<br />

2<br />

1<br />

0<br />

Figure 7: Spider diagram of modal scores for categories 5-6.<br />

6.4<br />

6.5<br />

6.3<br />

6.6<br />

6.2<br />

6.1<br />

388<br />

1.2<br />

1.6<br />

3.2<br />

3.6<br />

5.2<br />

5.6<br />

1.3<br />

1.5<br />

3.3<br />

3.5<br />

1.4<br />

3.4<br />

5.3<br />

5.5<br />

5.4


4. Analysis<br />

Ari Riabacke et al.<br />

The results differ markedly between categories of statements. In category 1, Visions, objectives, and<br />

strategies, there is good agreement for all except two statements. There seems to be a fairly good<br />

grounding of BI systems in the organisations’ control documents and visions, showing a conceptual<br />

integration of BI as decision support and with an understanding of how the system may provide support.<br />

1.2 and 1.6 did indicate the opposite, these two having modal scores of 3 and 2 respectively. Those two<br />

latter statements referred to specific management issues which were perceived to be less integrated with<br />

BI visions.<br />

In category 2, Business values from BI systems, there was also a good agreement with the statements in<br />

general. There is a clear indication that the needs are being catered for and that perceived business<br />

values are derived from business intelligence information. The only exception here is statement 2.3 which<br />

deals with the source of promotion for BI solutions, indicating that the need of a BI system is not primarily<br />

perceived by the benefiting organisation but rather promoted by external sources.<br />

In category 3, Requirements analysis and needs, all indicators show that there was good agreement with<br />

the statements. This implies that the most businesses indeed have a fairly good control over the design<br />

and implementation of BI systems from an organisational perspective.<br />

In category 4, Change management, there was much less agreement with the statements. In particular,<br />

statements 4.1 – 4.3 indicated stronger disagreement with the statements, implying problems with<br />

integrating the information in decision making processes. The result for statement 4.5 is clearly positive,<br />

dealing with user involvement in general, whereas 4.4 and 4.6 indicate a neutral to agreeing position with<br />

the statements. The latter three are statements more typical to information systems in general.<br />

In category 5, Technical solutions, most statements are agreeable for a majority of respondents. This<br />

category deals with technical issues on implementation as opposed to category 3 which deals with<br />

organisational issues. Like category 3, this is not overly complicated to deal with for accomplished<br />

organisations. There is, however, strong disagreement regarding statement 5.3, again dealing with top<br />

management. On the other hand, there is also disagreement among researchers on the involvement of<br />

top management in these issues in contrast with involvement in organisational issues which are<br />

considered much more pivotal to success.<br />

Finally, in category 6, Decision making support, we find the strongest area of disagreement, with all six<br />

indicators being disagreed with on an aggregate level. This clearly implies that the largest set of problems<br />

with BI systems lies in integrating them with decision making processes in the businesses, indicating<br />

problems with BI scope rather than the BI systems themselves.<br />

The same picture is painted by the modal scores, see Figures 5-7.<br />

5. Conclusion<br />

Of the six problem areas studied in this paper, clearly Change management and Decision making support<br />

are those that contribute most to BI systems not being as efficiently functional in an organisational sense<br />

as anticipated. The area Visions, objectives, strategies is mixed, with some problems, and Technical<br />

solutions, Business values from BI systems and Requirements analysis and needs are less problematic<br />

than the others. This indicates that the perceived value from a BI system does not lie in its decision<br />

support function and that there is little understanding in how to use it as decision support, i.e. a problem<br />

within the wider concept of BI scope.<br />

In other words, when looking back at the triangle in Figure 1, we have found that most efforts done<br />

towards the integration of business intelligence systems in organizations focuses on the right hand side<br />

of the triangle in Figure 1, i.e. on the technical parts of the BI solution.<br />

6. Discussion and outlook<br />

While almost every for-profit organization builds business cases or similar around new investments in<br />

general, this is much more seldom done for BI systems, where the inability to measure its impact renders<br />

standard measurement models unusable. There is of course a reason for BI systems winding up in the<br />

state they are. One hypothesis that need to be further examined in a more qualitative investigation is that<br />

this is a result of that BI systems are treated as just another kind of traditional information systems. In<br />

389


Ari Riabacke et al.<br />

contrast to operational information systems which directly support operational processes that often are<br />

rather well defined, BI systems have the capability to support follow-up and decision making processes<br />

which typically are less defined in an enterprise. Hence, for an organization, a BI system is<br />

complementary to an operational information system, serving as a support for managing the business at<br />

several different levels. Surely, decision processes could also be modelled using process modelling<br />

techniques, but not the same as operational business processes. However, when lacking another<br />

approach, it is understandable that BI systems are treated as “just the next information system” when it is<br />

being designed, implemented, and maintained. At present, the main objective when implementing a BI<br />

system seems to be that of simply getting it to operate, with too little emphasis on actually getting it to<br />

work within its context and thus provide value. This also supports the statement that the practice of BPM<br />

concepts is not widespread or mature in the responding organizations, or that the BI system is neglected<br />

in this practice. Therefore, too many organizations nowadays express a neutral or unhappy view on their<br />

BI systems. Avoiding this calls for another approach in the design phase, approaches stimulating how the<br />

information is going to be used and by whom, as opposed to approaches focusing on simply getting it<br />

implemented and usable.<br />

In order to develop good BI solutions that support organizations in reaching their overall goals, we must<br />

strive for a more holistic view on the implementation of BI systems, for instance by not implementing BI<br />

systems without having a proper treatment and understanding of the business processes and the related<br />

decision making activities. This means that the solutions must be based on knowledge from a number of<br />

areas in addition to database management and computer engineering, i.e. the balance between decisionmakers,<br />

business models and the technical support system must be better managed. In most current BI<br />

solutions there is a lack of balance in terms of focus on the different corners in the triangle. This calls for<br />

requirements specification and design processes focusing on and identifying the business and decision<br />

processes that are to benefit from the information provided through the system as well as the<br />

organizations overall strategy.<br />

7. Appendix<br />

7.1 The questionnaire<br />

The following three sub-sections show the statements in the questionnaire as they appear in the results<br />

section. They have been grouped into three clusters for presentational purposes.<br />

7.1.1 Categories 1-2<br />

1.1 The BI-systems support the work towards fulfilling the organizations objectives, strategies, and<br />

visions.<br />

1.2 The BI strategy is well established at the organization’s top management.<br />

1.3 BI is an important part of the organization’s strategy process.<br />

1.4 We have a clear understanding of how the BI systems will aid us in reaching stipulated<br />

objectives.<br />

1.5 We ensure that the technical solution, the organization, and the business objectives are<br />

consistent in our solution for BI.<br />

1.6 BI and BI-related questions are a common issue on the regular management meetings.<br />

2.1 The BI systems are tied to the needs of the organization.<br />

2.2 The BI systems provide information relevant to the different business units and their needs.<br />

2.3 It is clear who promote the needs for BI systems.<br />

2.4 The BI systems support the organization’s core processes.<br />

2.5 The development of our BI systems is solely driven by our business needs.<br />

2.6 We always create a “business case” prior to investments in BI systems.<br />

390


7.1.2 Categories 3-4<br />

Ari Riabacke et al.<br />

3.1 Our BI systems provide information that is easy to understand and use in decision making,<br />

3.2 The BI systems support the organization in reaching its overall business objectives.<br />

3.3 We have a clear distribution of roles and responsibilities when designing our BI systems.<br />

3.4 We have a clear distribution of roles and responsibilities when implementing our BI systems.<br />

3.5 The ownership of BI system implementation projects lies on the using part of the organization.<br />

3.6 Technical experts, users, and management are always involved in the design and implementation<br />

of BI systems.<br />

4.1 We secure that the foreseen users utilize the information provided by the BI systems.<br />

4.2 We have well established routines for how to prepare the organization and its users to use new<br />

technology and new information.<br />

4.3 We have clear ideas on how new available information will fit with existing decision processes.<br />

4.4 We allocate funds for education of personnel when investing in BI systems.<br />

4.5 We secure that the users are involved in development and implementation of BI systems.<br />

4.6 We secure that the organization and its individuals know how and why they are to use the BI<br />

systems and the information they provide.<br />

7.1.3 Categories 5-6<br />

5.1 We secure that the technical parts of the BI systems support stipulated business objectives.<br />

5.2 The business put clear requirements on what is desired from the technical parts of the BI<br />

systems.<br />

5.3 Corporate management is an active part in the technical requirements specification process for<br />

the BI systems.<br />

5.4 The users are an active part in the requirements specification process for the BI systems.<br />

5.5 There is a clear and structured user representation in the selection of technical solutions for our<br />

BI systems.<br />

5.6 The users of our BI systems is an active part when deciding on presentation formats of the<br />

information provided by the systems.<br />

6.1 We measure the business value generated by our BI systems in a quantitative fashion.<br />

6.2 There are clear processes for how the information provided by BI systems are be used in<br />

operational decision making.<br />

6.3 We use essentially all information accessible from our BI systems.<br />

6.4 There are clear guidelines and processes for how to cope with information anomalies or<br />

deviations from expectations.<br />

6.5 We utilize the information provided from our BI systems in a majority of the business decisions<br />

made in our company.<br />

6.6 We continuously secure that our BI systems provide relevant information for our business units.<br />

391


References<br />

Ari Riabacke et al.<br />

Bajaj, A. and Nidumolu, S. R. (1998) “A Feedback Model to Understand Information System Usage”, Information &<br />

Management, Vol. 33, pp. 213-224.<br />

Bogdana, P. I., Felicia, A., and Delia, B. (2009). “The Role of Business Intelligence in Business Performance<br />

Management”, Annals of Faculty of Economics, Vol. 4, No. 1, pp. 1025-1029.<br />

Bryman, A. (1989) Research Methods and Organizational Studies, Routledge, New York.<br />

French, S. and Ríos Insua, D. (2000) Statistical Decision Theory, Arnold, London, 2000.<br />

Frolich, M. N. and Ariyachandra, T. R. (2006). “Business Performance Management: One Truth”, Information<br />

Systems Management, Vol. 23, No. 1, pp. 41-48.<br />

Hannula, M. and Pirttimäki, V. (2003) “Business Intelligence Empirical Study on the top 50 Finnish Companies”,<br />

Journal of American Academy of Business, March 2003, pp. 593-599.<br />

Keen, P. G. W. (1981) “Value Analysis: Justifying Decision Support Systems”, MIS Quarterly, Vol. 5, No. 1, pp. 1-15.<br />

Lee, D., Newman, P. and Price, R. (1999) Decision Making in Organisations, Calays Ltd., Ives plc.<br />

Pomerol, J.-C. and Adam, F. (2004) “Practical Decision Making – From the Legacy of Herbert Simon to Decision<br />

Support Systems”, Proceedings of Decision Support in an Uncertain World: The IFIP TC8/WG8.3 International<br />

<strong>Conference</strong>, pp. 647-657.<br />

Rasmussen, J. (2000) “Human Factors in a Dynamic Information Society: Where are we Heading?”, Ergonomics,<br />

Vol. 43, No. 7, pp. 869-879.<br />

Sabherwal, R., Jeyaraj, A. and Chowa, C. (2006) “Information Systems Success: Individual and Organisational<br />

Determinants”, Management Science, Vol. 52, No. 12, pp. 1849-1864.<br />

Riabacke, A. (2006) “Managerial Decision Making Under Risk and Uncertainty”, IAENG International Journal of<br />

Computer Science, Vol 32., No. 4, pp. 453-459.<br />

Shim, J. P., Warkentin, M., Courtney, J. F., Power, D. J., Sharda, R. and Carlsson, C. (2002) “Past, Present and<br />

Future of Decision Support Technology”, Decision Support Systems, Vol. 33, No. 2, pp. 111-126.<br />

Speier, C. and Morris, M. G. (2000) “The Influence of Query Interface Design on Decision making Performance”, MIS<br />

Quarterly, Vol. 27, No. 3, pp. 397-423.<br />

Stenfors, S., Tanner, L., Syrjänen, M., Seppälä, T. and Haapalinna, I. (2007) “Executive Views Concerning Decision<br />

Support Tools”, <strong>European</strong> Journal of Operational Research, Vol. 181, pp. 929-938.<br />

Watson, H. J. and Wixom, B. H. (2007) “The Current State of Business Intelligence”, Computers, Vol. 40, No. 9, pp.<br />

96-99.<br />

Turban, E., Sharda, R., and Delen, D. (2010). Decision Support and Business Intelligence Systems, 9 th ed., Prentice<br />

Hall Press, Upper Saddle River NJ.<br />

Yin, R. K. (2003) Case Study Research: Design and Methods, 3 rd ed., Thousand Oaks: Sage Publications.<br />

392


Assessing Future Value of Investments in Security-Related IT<br />

Governance Control Objectives – Surveying IT Professionals<br />

Waldo Rocha Flores, Teodor Sommestad, Hannes Holm and Mathias Ekstedt<br />

Royal Institute of Technology, Stockholm, Sweden<br />

waldorf@ics.kth.se<br />

teodors@ics.kth.se<br />

hannes.holm@ics.kth.se<br />

mathiase@ics.kth.se<br />

Abstract: Optimizing investments in IT governance towards a better information security is an understudied topic in<br />

the academic literature. Further, collecting empirical evidence by surveying IT professionals on their relative opinion<br />

in this matter has not yet been explored to its full potential. This paper has tried to somewhat overcome this gap by<br />

surveying IT professionals on the expected future value from investments in security-related IT governance controls<br />

objectives. The paper has further investigated if there are any control objectives that provide more value than others<br />

and are therefore more beneficial to invest in. The Net Present Value (NPV) technique has been used to assess the<br />

IT professionals relative opinion on the generated future value of investments in 19 control objectives. The empirical<br />

data was collected through a survey distributed to professionals from the IT security, governance and/or assurance<br />

domain and analyzed using standard statistical tools. The results indicate that the vast majority of investments in<br />

control objectives is expected to yield a positive NPV, and are beneficial to an organization. This result implies that<br />

investments in control objectives are expected to yield positive Net Present Value for the firm, which is an important<br />

finding since many of the benefits from an investment are indirectly related an may occur well into the future. The<br />

paper moreover contributes in strengthening the link between IT governance and information security.<br />

Keywords: IT governance, control objectives, Information security, net present value<br />

1. Introduction<br />

Contemporary enterprises are largely dependent on Information Technology (IT) as it supports many<br />

critical business and administrative functions. This dependency has unfortunately led to an increase in<br />

potential threats to the enterprise and its information assets. Enterprises are therefore compelled to<br />

invest in information security to protect information assets, to avoid negative consequences that poor<br />

information security can lead to, and create value for their shareholders. Value for an enterprise’s<br />

shareholders is created by making beneficial real investment decisions. The meaning of real investments<br />

is expenditures that generate cash in the future and, as opposed to financial investments, like stocks and<br />

bonds, are not financial instruments that are traded in the financial markets (Grinblatt and Titman 1998).<br />

To aid a decision maker in determining whether an security investment create value for the shareholder a<br />

responsible manager need to argue and communicate the rational with the choice of the investment<br />

alternative (Johnson 2006). By focusing on communicating security issues concerning business issues<br />

and providing related business benefits, risks and benchmarks, the importance of a security investment<br />

can be heightened. Financial metrics such as net present value (NPV) improves the effectiveness of the<br />

communication and can therefore be used to evaluate different security investment alternatives so that<br />

the most beneficial investment alternative is identified (Theo, Renkemaa and Berghoutb 1997). The use<br />

of NPV, explicitly, as an evaluation tool for security investments at the proposal stage has also been<br />

suggested by (Gordon and Loes 2002).<br />

IT governance provides value by harmonizing decisions about the management and use of IT with<br />

desired business behaviors. This implies that IT governance investments pay off (Van Grembergen and<br />

De Haes 2008) (Weill and Ross 2004) (Simonsson, Johnson and Ekstedt 2010). A core objective for IT<br />

governance is information security (Calder and Watkins 2008), and it is therefore reasonable to believe<br />

that investments in IT governance also are beneficial for an organizations information security. A problem<br />

is, however, that benefits to the organization are often indirectly related to the investment and may occur<br />

well into the future. It may therefore be difficult for managers in an organization to demonstrate the<br />

expected future value of investing in control objectives. And to effectively direct investments endeavors, a<br />

crucial question is if there exist any differences in how effective these investments are in generating<br />

value. The need for decision support regarding investment opportunities in security is therefore needed.<br />

In this article we have investigated the value in terms of reduced negative consequences from security<br />

incidents generated from investments in IT governance control objectives (time, people or money etc.).<br />

393


Waldo Rocha Flores et al.<br />

To answer this question we have surveyed IT professionals in the security, governance and/or assurance<br />

domain and asked them on their opinion regarding the value of investments in IT governance control<br />

objectives. The Net Present Value (NPV) technique was used in order to assess the IT professional’s<br />

relative opinion on the value these investments is expected to provide.<br />

The structure of the paper is outlined as follows. Section 2 presents some related work to motivate our<br />

contribution in respect to previous research. In section 3 the control objectives from COBIT Security<br />

Baseline and the Net Present Value technique is presented. Section 4 presents the methodology used in<br />

the study. Section 5 presents and analyses the results. Section 6 discusses the results and finally,<br />

section 7 concludes the paper.<br />

2. Related works<br />

In the IT governance field there exists several studies on the impact of different components on overall<br />

goals. COBIT (ISACA 2007a) describes what processes that are primarily and secondarily related to<br />

several information criteria in its appendices. (Simonsson, Johnson and Ekstedt 2010) have assessed the<br />

impact of IT governance maturity on the external quality of delivered IT services experienced by the<br />

business. The authors further also investigated the relative importance of different IT governance<br />

processes. In addition, (Debreceny and Gray 2009) assessed the IT governance maturity of<br />

organizations with respect to predefined IT governance maturity attributes.<br />

The use of an information security standard as an IT governance framework has been proposed in<br />

(Calder and Watkins 2008). In this work, the authors discussed the relevance of the information security<br />

standard management standard ISO 27000 as a framework for fulfilling general objectives with IT<br />

governance.<br />

This study examines how control objectives associated with IT governance support information security<br />

and generate future value in terms of reducing negative consequences from security incidents. There are<br />

several studies that assess the importance of different security goals, for instance (Bartolini and Sallé<br />

2004) (Su, Bolzoni and Eck 2007) (Neubauer, Klemen and Biffl 2005). These authors have assessed the<br />

priority or impact of security objectives on information security. This article, however, assess the impact<br />

of investment in IT governance objectives on security.<br />

The impact of objectives related to security can be assessed in a number of ways. Economic methods<br />

are often used to evaluate different security strategies. For example, (Gordon and Loes 2002) (Mercuri<br />

2003) (Sonnenreich, Albanese and Stout 2006) have described methods where economic variables were<br />

used to evaluate security alternatives regarding the cost of investments and the value they provide.<br />

Security per se does not provide business value. The benefit of security investments are instead<br />

assessed in terms of the reduced cost from security incidents. Therefore, economic techniques such as<br />

the NPV or return on investment (relabeled as return on security investment) have been suggested by<br />

(Gordon and Loes 2002) (Sonnenreich, Albanese and Stout 2006).<br />

In the present paper, NPV technique is used to empirically assess the relative value investments in 19 IT<br />

governance control objectives are expected to provide. To the authors knowledge this has not yet been<br />

presented in the academic literature.<br />

3. COBIT security baseline and the Net Present Value<br />

This section describes the best-practice guideline that served as a basis for the survey and the evaluation<br />

technique used to evaluate generated future value of investment in control objectives.<br />

Control Objectives for Information and related Technology Standards (COBIT) is currently the most wellestablished<br />

best-practice guideline for IT governance. It outlines 34 IT processes, their purpose and the<br />

controls that should be used to govern them (ISACA 2007a)(Debreceny and Gray 2009) (Simonsson,<br />

Johnson and Ekstedt 2010). A core part of COBIT concerns information security and the 19 most<br />

essential IT governance control objectives toward better information security has therefore been<br />

extracted from COBIT and presented in COBIT Security Baseline (ISACA 2007b). In this paper COBIT<br />

security baseline (ISACA 2007b) was used as the basis for developing the survey used for collecting<br />

data.<br />

COBIT security baseline describes 19 security-related control objectives that are cross-referenced to<br />

related COBIT processes and related control sections in ISO/IEC 27002:2005 (ISO/IEC 2005). The<br />

394


Waldo Rocha Flores et al.<br />

control objectives are grouped into four domains: Plan and Organize (PO), Acquire and Implement (AI),<br />

Deliver and Support (DS), and Monitor and Evaluate (ME) (c.f. Figure 1). The domains map to IT´s<br />

traditional responsibility areas of plan, build, run and monitor. Within the COBIT framework, these<br />

domains, as shown in Figure 1, cover the following:<br />

Plan and Organize (PO): This domain provides direction to solution delivery and service delivery. It<br />

covers strategy and tactics, and concerns the identification of the way IT can best contribute to the<br />

achievement of the business objectives.<br />

Acquire and Implement (AI): This domain provides the solutions and passes them to be turned into<br />

services. To realize the IT strategy, IT solutions need to be identified, developed or acquired, as well<br />

as implemented and integrated into the business process. In addition, changes in and maintenance<br />

of existing systems are covered by this domain to make sure the solutions continue to meet business<br />

objectives.<br />

Deliver and Support (DS): This domain receives the solutions and makes them usable for end users.<br />

It is also concerned with the actual delivery of required services, which includes service delivery,<br />

management of security and continuity, service support for users, and management of data and<br />

operational facilities. All IT processes need to be regularly assessed over time for their quality and<br />

compliance with control requirements.<br />

Monitor and Evaluate (ME): This domain serves the purpose of monitoring all processes to ensure<br />

that the direction provided is followed. Further, the domain addresses performance management,<br />

monitoring of internal control, regulatory compliance and governance.<br />

Monitor and Evaluate<br />

17.Regularly monitor the performance of information<br />

security.<br />

18.Gain confidence and trust in security through<br />

reliable and independent sources.<br />

19.Ensure that information security functions comply<br />

with applicable laws, regulations and other external<br />

requirements.<br />

Deliver and Support<br />

11.Define and manage security aspects of service<br />

levels.<br />

12.Manage security aspects of third-party services.<br />

13.Ensure that the enterprise is capable of carrying on<br />

its day-to-day automated business activities with<br />

minimal interruption from a security incident.<br />

14.Ensure that all configuration items are<br />

appropriately secured and security risks minimized by<br />

ensuring the enterprise’s awareness of its IT-related<br />

assets and licenses.<br />

15.Ensure that all data remain complete, accurate and<br />

valid during input, processing, storage and<br />

distribution.<br />

16.Protect all IT equipment (hardware) from damage.<br />

IT Governance Control Objectives toward<br />

better information security<br />

Plan and Organize<br />

1.Identify information and services critical to the<br />

enterprise and consider their security requirements.<br />

2.Define and communicate information security<br />

responsibilities.<br />

3.Define and communicate management aims and<br />

directions with respect to information security.<br />

4.Ensure that security functions are staffed properly<br />

with people who possess the necessary skills to fulfill<br />

the role.<br />

5.Discover, prioritize, and either contain or accept<br />

relevant information security risks.<br />

Acquire and Implement<br />

6.Consider security when identifying, automated<br />

solutions.<br />

7.Consider security when acquiring and maintaining<br />

the technology infrastructure.<br />

8.Consider security when enabling operational use.<br />

9.Ensure that all changes, including patches, support<br />

enterprise objectives and are carried out.<br />

10.Ensure that all new systems and changes are<br />

accepted only after sufficient testing of security<br />

functions.<br />

Figure 1: IT governance control objectives drawn from COBIT security baseline<br />

The NPV method was used to assess the generated future value of the 19 control objectives. The starting<br />

point in the NPV method is the cost of capital for an investment, otherwise known as the required return.<br />

The cost of capital is the amount that an investor requires to compensate her for the time value of money<br />

tied up in the investment and for taking on risk in the investment. Thus, it represents the costs for her to<br />

take on the investment alternative and it is the minimum return that investors expect for providing capital<br />

395


Waldo Rocha Flores et al.<br />

to an investment, thus setting a benchmark that an investment has to meet. For an investment to add<br />

future value, it must earn more than the cost of capital. When calculating the NPV for an investment the<br />

following formula is used:<br />

NPV<br />

∞<br />

∑<br />

Rt<br />

t=<br />

( 1+<br />

i<br />

= t<br />

0 )<br />

The cost of capital rate ( ) is used as the discount rate. This is the rate of return that could be earned on<br />

an alternative investment with similar risk. is the net cash flow, i.e. the amount of cash, inflow minus<br />

outflow, at time To calculate the NPV for an investment each cash inflow/outflow associated with it is<br />

discounted back to its present value. The NPV is the sum of all present values from the time series of<br />

cash flows. If the NPV is larger than zero it is beneficial to make the investment and the investment is<br />

therefore a good real investment opportunity. If NPV is less than zero it will not add value and thus be a<br />

less attractive investment opportunity. Applied to our case a good real investment opportunity is an<br />

investment in security that generates future value in terms of reduced impacts from security incidents and<br />

provides value to the shareholder. To estimate the value of investments in control objectives the NPV of<br />

all 19 control objectives was assessed by surveying IT professionals.<br />

4. Method<br />

This study utilizes a survey as a measurement tool, this due to the obvious strengths in terms of<br />

statistical analysis and cost efficiency. The aim of the study is not to reach in-depth information regarding<br />

each of the 19 control objectives; it is simply to gain an understanding of the relative impact of control<br />

objectives on information security and can thus be categorized as an exploratory study.<br />

4.1 Population and sample<br />

When constructing a survey it is important to specify a population with favorable attributes and choose a<br />

representative sample from that population (Saunders, Lewis and Thornhill 2009). A natural prerequisite<br />

for respondents is that they can relate to the questions of the survey (Blair 2005). In this case it meant a<br />

population of experienced professionals in the IT security, governance and/or assurance domain.<br />

A strategic sample consisting of a conference with approximately 100 participants satisfying the<br />

population constraints was used. Of this population 22 answered the survey. 15 respondents were<br />

professionals in IT Security, 9 were professionals in IT Assurance and 11 were professionals in IT<br />

governance. The most common combination was IT Security and IT governance. Four respondents came<br />

from the banking industry, six from the energy industry, three from the public administration sector, two<br />

from the telecommunication sector, two from transport logistics and one consultant. One respondent did<br />

not specify his or her sector.<br />

The majority of the respondents were CISA (Certified Information System Auditor) and/or CISM (Certified<br />

Information Security Manager); twelve respondents were CISA, six CISM and three CGEIT (Certified in<br />

the Governance of Enterprise IT). Only four respondents did not hold any certification. Thus, the<br />

respondents originate from many different branches and a diverse number of roles. Although slightly<br />

diminutive this sample is considered representative of the population.<br />

4.2 The survey<br />

A face-to-face survey was carried out since all respondents were at the same geographical location (Blair<br />

2005). The survey consisted of two pages of which the first according to recommendation by (Blair 2005)<br />

introduces the concepts of COBIT Security baseline, NPV, and a description of how to answer the<br />

questions. Furthermore, the first page also includes three questions used to assess background<br />

information of respondents. The first question concerned the domain the IT professional work in, the<br />

second if the respondent holds any certification and the third in which industry the IT professionals are<br />

active in. The second page of the survey consisted of 19 questions utilized in order to gain information<br />

regarding the significance of the IT governance control objectives in the COBIT security baseline. All of<br />

the 19 questions included in the survey were taken directly from COBIT security baseline without any<br />

manipulation.<br />

396<br />

(1)


Waldo Rocha Flores et al.<br />

For each of the 19 control objectives the respondents were asked to indicate the size of the<br />

corresponding NPV on a scale. In figure 2, the 5 questions belonging to the Plan and Organize domain<br />

are presented to provide information of the question format. The questions were answered using ratio<br />

scale, with 0 NPV in the middle, positive NPV to the right and negative NPV to the left. The results were<br />

interpreted using a ruler with a millimeter scale, rounded to the nearest complete mm. The outcome was<br />

then rescaled to the interval [-5,+5] in order to generate more pedagogical results.<br />

Please estimate how investments in each of the areas below would add additional value (NPV) to an organization in<br />

the industry you work in.<br />

4. Control objectives (investment opportunities)<br />

Yields the following NPV<br />

Negative Positive<br />

4.1. Identify information and services critical to the enterprise and<br />

consider their security requirements.<br />

4.2. Define and communicate information security responsibilities.<br />

4.3. Define and communicate management aims and directions<br />

with respect to information security.<br />

4.4. Ensure that security functions are staffed properly with people<br />

who possess the necessary skills to fulfill the role.<br />

4.5. Discover, prioritize, and either contain or accept relevant<br />

information security risks.<br />

Figure 2: Question format<br />

4.3 Analysis methods<br />

Three tools were utilized to analyze the results of the survey: box plots, tests for normality and statistical<br />

measurements.<br />

Box plots (sometimes referred to as “box and whisker plots”) were first used to evaluate the results. Box<br />

plots were chosen to be used due to the favorable non-parametric distribution requirement by the tool.<br />

The box plot provides a graphical representation of a dataset according to percentiles. The bottom of a<br />

box (the lower quartile, Q1) is the 25 th percentile and the top of the box (the upper quartile, Q3) is the 75 th<br />

percentile. The band in the box represents the median (50 th percentile). Data in the 1 st -24 th percentile and<br />

the 76 th -100 th percentile are represented by whiskers and/or dots (outliers). More information regarding<br />

box plots can be found in (McGill, Tukey och Larsen 1978).<br />

Tests for normality were performed using QQ-plots (Warner 2008). Outliers who did not meet the<br />

normality assumption were removed according to the recommendation by (Montgomery 2005) in order to<br />

improve the overall data quality. The evaluation of the data set with box plots gave an understanding of<br />

the data before manipulation; therefore test for normality was performed after analyzing the box plots.<br />

Standard statistical measurements such as mean, standard deviation and confidence intervals were<br />

finally used to present the survey results. Descriptions of these measures can be found in (Montgomery<br />

2005).<br />

5. Results and analysis<br />

A control objective with a positive NPV should be interpreted as a control objective that adds value and is<br />

beneficial to invest in. If the NPV is less than zero it will not add value and is therefore not an attractive<br />

investment opportunity. Further, the control objective with highest NPV will be the most attractive<br />

investment opportunity when compared to other investments in control objectives with similar risk. Figure<br />

3 shows a box plot over the 19 control objectives and table 1 describes their mean, standard deviation<br />

and confidence intervals. The results from the survey indicate that the vast majority of security<br />

investments have a positive net present value, and are thus beneficial for an organization to invest in.<br />

The mean of all NPVs but one is positive. A 95 percent confidence interval over the mean values of<br />

respondent assessments indicate that the true mean (the mean of the population) lies between the upper<br />

397


Waldo Rocha Flores et al.<br />

(UB) and lower (LB) in the 95 % of the cases. In the remaining part of this section, the results will be<br />

discussed on a domain-level and thereafter each domain will be discussed.<br />

Figure 3: Assessments of control objectives investment potential<br />

Table 1: Control objectives ranking<br />

Objective Mean Std dev LB (95%) UB (95%)<br />

CO1 2.0 1.4 1.4 2.6<br />

CO2 1.5 1.3 1.0 2.1<br />

CO3 1.7 1.5 1.1 2.3<br />

CO4 1.1 1.5 0.5 1.7<br />

CO5 2.0 1.1 1.5 2.5<br />

CO6 1.2 1.3 0.6 1.7<br />

CO7 1.3 1.3 0.8 1.9<br />

CO8 1.3 1.2 0.8 1.9<br />

CO9 1.6 1.5 1.0 2.3<br />

CO10 1.5 1.6 0.8 2.2<br />

CO11 0.5 1.8 -0.3 1.2<br />

CO12 1.0 2.2 0.0 1.9<br />

CO13 1.6 1.8 0.8 2.4<br />

CO14 0.8 1.6 0.1 1.5<br />

CO15 1.3 1.7 0.5 2.0<br />

CO16 -0.2 1.9 -1.0 0.7<br />

CO17 0.9 1.7 0.1 1.6<br />

CO18 0.1 1.2 -0.4 0.6<br />

CO19 1.3 1.5 0.6 1.9<br />

Mean 1.2 1.5 0.5 1.9<br />

398


5.1 Domains<br />

Waldo Rocha Flores et al.<br />

A box plot of the results aggregated to the domain level can be seen in Figure 4. Plan and organize (PO,<br />

CO1-CO5) have the highest median, followed by acquire and implement (AI, CO6-CO10). Deliver and<br />

support (DS, CO11-CO16) and monitor and evaluate (ME, CO17-CO19) have relatively similar medians<br />

and ranges, although DS have a higher degree of spread of data. Table 2 shows the mean of the domain<br />

experts’ assessments of NPV for control objectives within each domain. As can be seen here the highest<br />

mean is associated to the domain Plan and organize and the lowest is associated with Deliver and<br />

Support and Monitor and evaluate. Also, the mean of the NPV is positive for all four domains.<br />

Figure 4: Box plots per domain<br />

Table 2: Control domains mean and spread<br />

Domain Mean Std dev LB (95%) UB (95%)<br />

PO 1.6 1.1 1.2 2.1<br />

AI 1.3 1.1 0.8 1.8<br />

DS 0.8 1.3 0.3 1.4<br />

ME 0.8 1.0 0.3 1.2<br />

5.1.1 Plan and organize<br />

As can be seen in the box plot in Figure 3, investments in all control objectives in this domain are<br />

presumed to yield relatively high NPVs. QQ-plots indicate that the outliers need to be removed in order to<br />

satisfy the assumption of normality. The variables with the highest median, smallest box and lowest<br />

range is CO1 (“Identify information and services critical to the enterprise and consider their security<br />

requirements”) and CO5 (“Discover, prioritize, and either contain or accept relevant information security<br />

risks”).<br />

5.1.2 Acquire and implement<br />

Investments in control objectives within this domain are generally believed to render a positive NPV. The<br />

two control objectives with highest means are CO9 (“Ensure that all changes, including patches, support<br />

enterprise objectives and are carried out”) and CO10 (“Ensure that all new systems and changes are<br />

accepted only after sufficient testing of security functions”) .The boxes in this domain are approximately<br />

the same size, with the exception of which is slightly larger than the rest. CO10 does however have a<br />

strong outlier on the negative side which increases its standard deviation (cf. figure 3 and table 1).<br />

399


5.1.3 Deliver and support<br />

Waldo Rocha Flores et al.<br />

There is some variation between the expected NPV of investments in this domain. For instance, the<br />

mean NPV associated with CO16 (“Protect all IT equipment from damage”) is negative, but CO13<br />

(“Ensure that the enterprise is capable of carrying on its day-to-day automated business activities with<br />

minimal interruption from a security incident”) has the fourth largest mean NPV. Compared to the other<br />

domain there seems to be a higher disagreement among respondents on the NPV associated with<br />

investments in this domain. The assessments of NPV associated with CO12 (“Manage security aspects<br />

of third-party services”) is for example spread between highest return assessed to well below zero.<br />

5.1.4 Monitor and evaluate<br />

The monitor and evaluate domain consist of three control objectives. Their mean values are associated<br />

with a comparably low NPV. For instance is CO18 (“Gain confidence and trust in security through reliable<br />

and independent sources“) assessed to yield a NPV close to zero. Also within this domain is the spread<br />

of assessments comparably high. A 95 percent confidence interval over the respondents’ assessments<br />

does for example stretch between 0.1 and 1.6 for CO17 (“Regularly monitor the performance of<br />

information security”).<br />

6. Discussion<br />

The result from this survey is intended to reflect the opinion from a number of professionals within the IT<br />

security, governance and/or assurance domain. The reliability and validity of the instrument used to<br />

measure their opinion is described in section 6.2. The section below will discuss the results of this<br />

measurement.<br />

6.1 Value of investments in control objectives<br />

This paper has provided empirical data from a panel of IT professionals regarding the value in terms of<br />

reduced negative consequences from security incidents generated from investments in IT governance<br />

control objectives (time, people or money etc.). 16 out of 19 control objectives are associated with a 95<br />

percent confidence interval that only span over the positive side of the scale. This indicates that the panel<br />

agrees that investments in IT governance control objectives strengthen security objectives and are<br />

therefore beneficial. This is a finding that can be useful for a practitioner as it may be difficult to<br />

demonstrate the expected NPV of investing in control objectives since the benefits to the firm are often<br />

indirectly related to the investment and may occur well into the future.<br />

When it comes to individual control objectives the results gives some indications that C05 (“Discover,<br />

prioritize, and either contain or accept relevant information security risks”) and CO1 (“Identify information<br />

and services critical to the enterprise and consider their security requirements”) are beneficial to engage<br />

in. The less beneficial control objectives are CO16 (“Protect all IT equipment from damage.”) and CO18<br />

(“Gain confidence and trust in security through reliable and independent sources”). The respondents’<br />

answers regarding individual control objectives vary, the box plot in Figure 3 and the 95 percent<br />

confidence interval in Table 1 indicate this. However, the variations are small. Two explanations for these<br />

small variations are possible:<br />

Positive lopsidedness<br />

Respondent experience.<br />

These are now further discussed.<br />

Positive lopsidedness: The mean of the respondents assessments result in a positive NPV for all control<br />

objectives but one. This reflects a general belief among respondents in the benefit associated with<br />

investments on control objectives associated with security. As earlier described the respondents are<br />

individuals which have a vested interest in IT /security governance and practice such activities on a daily<br />

basis. With this as a basis the positive lopsidedness in NPVs could be attributed to an expected bias in<br />

the respondent’s judgment. While the positive lopsidedness could be an effect of the respondents biased<br />

opinions it could also reflect the situation in enterprises today. The software security market has grown in<br />

recent years (Gartner 2009) (Gartner 2010) and forecasts of the security market such as (DefenseNews<br />

2010) predict an increase in coming years. Such data indicate that many organizations do see a benefit<br />

in investments that strengthen security objectives.<br />

400


Waldo Rocha Flores et al.<br />

Respondent experience: The respondents have different backgrounds and experiences. Some variation<br />

among their opinions can therefore be expected. The assessments made of NPV potentials were also<br />

made for organizations in the industry where the respondent is a practitioner. Some variation can thus be<br />

explained by differences among industries. For example, the regulations that apply to organizations vary<br />

with industry. A regulation might require certain investments to be made and in that way influence the<br />

NPV that can be obtained from future investments associated with a control objective.<br />

Based on the spread in this data it appears as the accuracy of IT professionals assessments of individual<br />

investment opportunities in the IT governance domain is associated with a significant uncertainty. Future<br />

work could investigate the reasons for the spread among the respondents’ rankings of such investment<br />

opportunities.<br />

6.2 Instrument validity<br />

This section first address the instrument’s validity in terms of: content-, construct-, external- and internal<br />

validity (Boudreau, Gefen and Straub 2001) (Brewer 2000).<br />

Content validity is the degree to which items in an instrument reflect the content universe to which the<br />

instrument will be generalized (Boudreau, Gefen and Straub 2001). E.g., if this study aimed to draw<br />

conclusions on profitability of the COBIT security baseline using only results from control objectives 1-4<br />

the content validity would be low. Normal methods of measuring content validity are through literature<br />

review and expert panels/judges (Boudreau, Gefen and Straub 2001). All control objectives of the COBIT<br />

security baseline are included in the survey. Furthermore, all survey questions are taken directly from the<br />

framework without any manipulation. However, while the baseline concepts are thoroughly covered there<br />

might be slight problems with the prioritization technique employed (NPV). The NPV of a potential<br />

investment is a commonly used indicator on how lucrative investments are for an organization, but when<br />

making investment decisions other factors could also be considered. In particular, the risk of the<br />

investment and the opportunity costs associated with it could be of essence. These factors are only<br />

implicitly measured in this survey and one should therefore be somewhat careful when interpreting the<br />

relative ranking associated with different investments. Finally, the survey was studied by two academic<br />

experts in the area who assessed the tools content validity as high.<br />

Construct validity is the extent to which an operationalization measures the concepts that it purports to<br />

measure (Boudreau, Gefen and Straub 2001), i.e. whether the survey measures the benefit of investing<br />

in different control objectives and domains. First of all, neither the COBIT security baseline nor<br />

prioritizations of its elements are abstract in such a manner that could cause significant problems with the<br />

construct validity. Second, the baseline and prioritization potency of NPV are generally thought of as<br />

highly valid theory. Finally, the tool was tested (and discussed) by two academic experts and one<br />

respondent which were thought to be representative of the specified population. These pilot studies both<br />

hinted toward high construct validity.<br />

Another important validation concept is external validity; the degree to which the results of a study can be<br />

generalized [30]. This study is on one hand built upon a sanctioned theoretical framework and a to some<br />

extent representative sample. However, it is to the authors’ knowledge also the first study performed<br />

relating to the COBIT security baseline and NPV; therefore it is not possible to compare the results of this<br />

study to the results of previous research. Also, the sample is too small to draw any certain general<br />

conclusions.<br />

An additional well regarded attribute of validity is internal validity (Brewer 2000). However, since no<br />

evaluation of causal relationships was made in this study, internal validity is not applicable here.<br />

7. Conclusion<br />

Optimizing the investments of resources in information security is a complex task. The study presented in<br />

this article has tried to somewhat make the investment decision more rational by investigating the value<br />

in terms of reduced negative consequences from security incidents generated from investments in IT<br />

governance control objectives (time, people or money etc.). By surveying IT professionals in the security,<br />

governance and/or assurance domain regarding the NPV that could be obtained from these investments<br />

the study found that investments in control objectives are beneficial for a firm to engage in. The paper<br />

has therefore provided data that supports the theory that investments in IT governance also are beneficial<br />

for the information security in an organization.<br />

401


Waldo Rocha Flores et al.<br />

The mean NPV vary between the control objectives, and there are some indications of control objectives<br />

that provide more value. However, this study cannot assert which control objective that is most beneficial<br />

to invest in. Interesting future work includes therefore more research into the impact of prioritization<br />

between control objectives. Research could investigate if the provided benefit for a firm depends on<br />

proper prioritization, and if so, identify which control objective/objectives that should be prioritized.<br />

References<br />

Bartolini, C, and M Sallé. "Business Driven Prioritization of Service Incidents." DSOM, LNCS 3248. 2004. 64-75.<br />

Blair, Czaja R. Designing surveys. Sage Publications Inc, 2005.<br />

Boudreau, Marie-Claude, David Gefen, and Detmar W Straub. "Validation in Information Systems Research: A Stateof-the-Art<br />

Assessment." MIS Quarterly 25, no. 1 (2001): 1-16.<br />

Brewer, Marilynn B. "Research Design and Issues of Validity." In Handbook of research methods in social and<br />

personality psychology, by Harry T Reis and Charles M Judd, 3-17. Cambridge: Cambridge University Press,<br />

2000.<br />

Calder, A, and S Watkins. IT governance A manager´s guide to Data Security and ISO 27001/ISO 27002. Kogan<br />

Page, 2008.<br />

Cronobach, L J. "Coefficient alpha and the internal structure of tests." Psychometrika, ([14] L.J. Cronbach,<br />

"Coefficient alpha and the internal structure of tests," Psychometrika, vol. 16, 1951, pp. 297-334.) 16 (1951):<br />

297-334.<br />

Debreceny, R, and G.L Gray. "IT Governance and Process Maturity: a Field Study." In Proceedings of the 42nd<br />

Hawaii International <strong>Conference</strong> on System Sciences. 2009.<br />

DefenseNews. IT Security Market Heats Up - Demand Growing for Guidance on Cyber Protection. DefenseNews,<br />

2010.<br />

Gartner. Gartner says worldwide security software market on pace to grow 8 per cent in 2009. Gartner, 2010.<br />

Gartner. Gartner says worldwide security software revenue grew 18.6 per cent in 2008. Gartner, 2009.<br />

Gordon, L.A, and M.P Loes. "The Economics of Information Security Investment." ACM Transactions on Information<br />

and System Security Vol. 5, No. 4, 2002: 438–457.<br />

Grinblatt, M, and S Titman. Financial Markets and Corporate Strategy. McGraw-Hill International Edition, 1998.<br />

Hardy, G. "Using IT governance and COBIT to deliver value with IT and respond to legal, regulatory and compliance<br />

challenges." information security technical report, 2006: 55-61.<br />

ISACA. COBIT Security Baseline: An Information Security Survival Kit 2nd edition. ISACA, 2007b.<br />

ISACA. Control Objective for Information and Related Technology Standards. ISACA, 2007a.<br />

ISO/IEC. "Code of Practice for Information Security management, ISO/IEC 27002:2005." Switzerland, 2005.<br />

Johnson, E.C. "Security awareness: switch to a better program." Network Security, 2006: 15-18 .<br />

McGill, Robert, John W Tukey, and Wayne A Larsen. "Variations of Box Plots." The American Statistician 32, no. 1<br />

(1978): 12-16.<br />

Mercuri, R.T. "Analyzing Security Costs." Communications of the ACM Vol. 46, No. 6, 2003: 15-18.<br />

Montgomery, Douglas C. Introduction to Statistical Quality Control. Vol. 5. New Jersey: John Wiley & Sons, Inc.,<br />

2005.<br />

Neubauer, T, M Klemen, and S Biffl. "Business Process-based Valuation of IT security." EDSER. St. Louis, Missouri,<br />

USA, 2005.<br />

Saunders, Mark, Philip Lewis, and Adrian Thornhill. Research Methods for Business Students. Edinburgh: Pearson<br />

Education Limited, 2009.<br />

Simonsson, M, P Johnson, and M Ekstedt. "The effect of IT Governance Maturity on IT Governance Performance."<br />

Information Systems Management, December 2010: 10-24.<br />

Sonnenreich, W, J Albanese, and B Stout. "Return On Security Investment (ROSI) - A practical Quantitative Model."<br />

Journal of Research and practice in Information Technology, Vol. 38, No. 1, 2006.<br />

Su, X, D Bolzoni, and P.V Eck. "Understanding and Specifying Information Security Needs to support the Delivery of<br />

High Quality Security Services." The international <strong>Conference</strong> on Emerging Security Information, Systems and<br />

Technologies. 2007.<br />

Theo, J.W, E.W Renkemaa, and Berghoutb. "Methodologies for information systems investments evaluation at the<br />

proposals stage: a comparative review." Information and Software Technology Vol. 39 No. 1, 1997: 1-13.<br />

Van Grembergen, W, and S De Haes. Implementing information technology governance: Models, practices, and<br />

cases. IGI Pub., 2008.<br />

Warner, R. M. Applied Statistics. SAGE Publications, 2008.<br />

Weill, P, and J.W Ross. IT governance: How top performers manage IT decision rights for superior results. Harvard<br />

Business School Press, 2004.<br />

Whitman, M.E. "In defense of the realm: understanding the threaths to information security." International Journal of<br />

Information Management, February 2004: 43-57.<br />

Yin, R K. Case study research: Design and Methods. Washington D.C: SAGE Publications, 2003.<br />

402


Evaluation Information Extraction for Health Text Categories<br />

Using C4.5 and Naïve Bayes<br />

Klaokanlaya Silachan and Panjai Tantasanawong<br />

Silpakorn University, Thailand<br />

Klao_99@yahoo.com<br />

Panjai@su.ac.th<br />

Abstract: Healthcare information technology, HIT or HICT supports health information seeking tasks. Data mining is<br />

an information extraction technique for identifying relevant information, structuring this information and subsequently<br />

providing the means to add textual information through mining the natural language. In the healthcare field,<br />

techniques summarise patient records, extract terms, ascertain new knowledge and provide diagnostics for various<br />

conditions. In this study we are interested in health information particularly with regard to the relationship between<br />

symptoms and cause in cardiovascular disease (CVD). Data filtered through a discovery pattern can be used for<br />

many tasks, including mining and evaluation to the accuracy of information extraction. In this paper an approach and<br />

framework of classification analysis has been proposed for evaluating health information from personal and medical<br />

data extracted from records and documents. The information extraction, as a feature of the selection from textual<br />

data, is significant in that it uses sliding window and TF-IDF techniques for a multiword check as well as for statistical<br />

matrix term weighting and rate of recurrence. Diagnosis of a probable cardiovascular disease is done by looking at<br />

the symptoms and mined data by way of the ML algorithm. The outcome of the specific medical conditions is<br />

considered and determined the probability of cardiovascular disease in order to classify the documents and<br />

cardiovascular disease by the use of rule induction. This model generates a classifier from the patient medical data<br />

training set. Two machine learning techniques, C4.5 and Naïve Bayes, have been applied where the domain concept<br />

is ‘diseases’. Evaluation of the accuracy for the disease term and document classification process for our theoretical<br />

patient’s symptom and diagnosis were measured precisely using recall and F-measure for comparison as well as to<br />

demonstrate the process of learning. The results show that in predicting the categories of cardiovascular disease, the<br />

C4.5 classifier has 99.8% rate of prediction accuracy (percentage correct) and performed better performance than<br />

the Naive Bayes.<br />

Keywords: information extraction, health text categories, C4.5, Naïve Bayes, evaluation<br />

1. Introduction<br />

The main goal of text mining is to enable the user to extract relevant information from textual resources<br />

and manage operations such as classification and summary support (Dalila 2004). Healthcare<br />

information technology, also known as HIT or HICT, extracts health information through a data mining<br />

technique that aims to identify relevant information, structure the information, and at the same time<br />

provide a means to allow the addition of semantics. It does this by processing the natural language and<br />

looking for patterns in unstructured text. Information extraction(IE) for healthcare has been designed to<br />

summarise patients’ medical records by extracting terms or call terms and then identify emerging patterns<br />

in information from the text on symptoms, conditions and diagnoses. The patient’s medical data are<br />

largely maintained in free-text form. Thus, using information extraction techniques as a reliable and<br />

efficient method to extract structured information for further data mining form, this free-text may greatly<br />

benefit research endeavours.<br />

This study is focused in health information and particularly the relationship between symptoms and<br />

causes in cardiovascular diseases (CVD). These disorders of the heart and blood vessels includes<br />

coronary heart disease (heart attacks), cardiovascular disease (stroke), raised blood pressure<br />

(hypertension), peripheral artery disease, rheumatic heart disease, congenital heart disease and heart<br />

failure. The major causes of cardiovascular disease are excessive smoking, physical inactivity, and an<br />

unhealthy diet, resulting in severe illness, disability and death. Data extraction has important implications<br />

for the classification of documented information. (K.S. Kavitha et al. 2010 and Reddy 2007). The<br />

probability of cardiovascular disease is based on symptoms diagnosis. For classification of diseases and<br />

medical document, cardiovascular medical data is obtained from the extracted text. Then the diagnosis of<br />

disease and classification will be evaluated for accuracy.<br />

It is necessary to measure the effectiveness of classifying data sets, medical data, for more precision and<br />

predictive accuracy in the classification of cardiovascular disease. In this paper an approach and<br />

framework is proposed that focuses on the evaluation of health information for the classification and<br />

analysis of cardiovascular disease case studies. The disease data can be extracted from a patient’s<br />

medical records and other documents. This step uses sliding windows and Term Frequency-inverse<br />

403


Klaokanlaya Silachan and Panjai Tantasanawong<br />

Document (TF-IDF) techniques for a multiword check as well as for statistical matrix term weighting and<br />

incidence in information extraction with the patient’s textual medical term data. The next step involves<br />

diagnosis of a probable cardiovascular disease by looking at the symptoms and mining the databases for<br />

classification with an ML algorithm. This model generates a classifier from the medical data training set<br />

based on the characteristics of texts already classified. Two machine learning techniques, C4.5 and<br />

Naïve Bayes, can be applied where the domain concept is ‘diseases’. Finally, the evaluation of accuracy<br />

of the classified disease for a theoretical patient treatment and the subsequent diagnosis can be<br />

measured precisely using recall and F-measure with an alternate machine learning algorithm (ML) for<br />

comparison and to demonstrate the process of learning.<br />

2. Related work<br />

Many researchers have studied text categorisation, medical health information and evaluation, such as<br />

Maria Grineva, et al. 2009, proposed a method for the extraction of key terms from textual documents. In<br />

this study the method is based on a combination of the following two techniques: a Wikipedia-based<br />

semantic relatedness measure of terms and an algorithm for detecting community structure of a network.<br />

Parikshit Sondhi et al. 2010 studied a shallow information extraction problem that involved extracting<br />

sentences of a given set of topic categories from medical forum data. Given a corpus of medical forum<br />

documents, which the research goal is to extract two related types of sentences that describe a<br />

biomedical case, i.e.: medical problem descriptions and medical treatment descriptions. Such an<br />

extraction task directly generates medical case descriptions that can be useful in many applications.<br />

Experimental results show that the accuracy was up to 75%.<br />

The Dittapol et al. 2009 study involved the application of methods of knowledge discovery from<br />

databases using decision tree algorithms for respiratory system diagnosis to classify patients from the<br />

Pranakornsri Ayudthaya Hospital into three groups. From the hospital’s medical records the study<br />

methods were knowledge discovery with the employment of ID3, C4.5 and CART decision tree<br />

algorithms and determination of the effectiveness of the three algorithms.<br />

Keku E 2005 proposed that trained interviewers should conduct annual follow up telephone interviews to<br />

ascertain any significant health events, including subsequent diagnostic tests, hospitalisations or death.<br />

Information on cohort hospitalisations and deaths is transmitted to the medical record abstraction unit<br />

who review death certificates and hospital records. Interviews with the next of kin and completed<br />

questionnaires by physicians and medical examiners or coroners were used to obtain information on<br />

deaths within the cohort.<br />

3. Methodology<br />

The proposed framework and method for information medical text extraction data for text term<br />

classification is explained in the diagram, Figure 1, below.<br />

Figure 1: Framework for information extraction and evaluation for health categories<br />

404


Klaokanlaya Silachan and Panjai Tantasanawong<br />

Textual data was extracted to classify the cardiovascular disease using the principles of extract term,<br />

sliding window and TF-IDF for matrix term weight and frequency. Groups of cardiovascular patients were<br />

classified using rule to recognize the relationship between the symptoms of the patient in the case where<br />

the conditions were symptomatic of cardiovascular disease. This Healthcare base was also used for<br />

building the cardiovascular diagnostic system. The features of C4.5 were applied to classify<br />

cardiovascular disease and relevant documents. Details of these steps include:<br />

3.1 Data set<br />

The construction of a corpus of medical text samples were collected from clinics and referenced from<br />

websites, such as the World Health Organization (WHO), and others given that a standard data collection<br />

should be used in the experimental evaluation. The corpus contains 300 cases. This text corpus is useful<br />

for training and testing with NPL and data mining purposes.<br />

3.2 Pre-processing<br />

In this step the data was checked for missing and incorrect values. Some of the data was alphanumeric.<br />

We found some inconsistencies in the patient’s records which we corrected manually by taking an<br />

external reference.<br />

3.3 Information extraction<br />

To understand the relationships between a patient’s cardiovascular symptoms and their treatment, we<br />

used the health service knowledge database and extracted those terms describing symptoms commonly<br />

found in the texts of patients’ medical records. Extraction of the symptom terms for classification was by<br />

single word or in phrases. As a result, a sliding window technique was used to retrieve phrases and then<br />

match it to a medical dictionary to control the vocabulary of the health information as well as to classify<br />

the type of heart diagnosis. (Dalila et al. 2004; Zhi-Hua etc al. 2003). The process is shown below.<br />

Sliding window for Information Extraction (IE) has the following steps:<br />

1. Extract key term with the number stored in the wordlist.<br />

2. Multiword created using sliding windows<br />

If(word != “ “)<br />

Word += “ “<br />

Word += WordList[j];<br />

3. Check for word or multiword.<br />

4. Count the frequency of the words.<br />

5. Check the word.<br />

6. Stored number of terms found.<br />

3.4 Feature selection<br />

The main idea of feature selection is to select subsets of features from the original documents. It is<br />

performed by keeping the highest score on words according to a predetermined measure of their<br />

importance. The term frequency-inverse document frequency (TF-IDF) is the ability to separate the words<br />

in the document. The frequency of words in the document and the ratio of the number of documents<br />

containing a particular word appear as an equation (Aurangzeb et al. 2010) and help evaluate how<br />

important a term is in proportion to the number of times a term appears in the document, which is then<br />

divided by the frequency of that word in the corpus (T.Gruber 1995; S.Izumi 2007). A terms weighting<br />

was calculated as shown:<br />

405


Klaokanlaya Silachan and Panjai Tantasanawong<br />

W(fd) = tf(fd) log (D / df(f)) (1)<br />

freqij<br />

tf(f i,dj) = (2)<br />

Max k freq Kj<br />

D<br />

idf = log (3)<br />

#(f 1 )<br />

Where, Term frequency (TF), invert Document Frequency(IDF) , W(fd) feature f d (current transaction),<br />

Tf(fd) feature f d normalize, feature d), D (training set), Df(f) feature f.<br />

Disease term weight-frequencies based on TF-IDF has the following steps:<br />

1. Store of words and the frequency in the table.<br />

2. Loop all terms found in each document.<br />

3. Check all terms found in each document.<br />

4. Check for this term in any other documents.<br />

5 Show the number of documents in which these words are found.<br />

6 Calculate values Log2 (N / nj).<br />

7. Loop all documents.<br />

8. Calculate frequency in all documents.<br />

9. Check for term found in each document or not.<br />

10. Calculate TF-IDF.<br />

11. Check store weight value.<br />

3.5 Medical data<br />

When IE is conducted from a medical text and the data is stored in a relational database as medical data.<br />

It consists of symptom words and phrases including terminology and conditions. The next step is to focus<br />

on analyzing the medical data alongside medical diagnostic data using classification rules in data mining<br />

to ascertain patterns of terms and to substantiate classification of diseases. Medical diagnostic data<br />

ought to be used in the experimental evaluation.<br />

3.6 Classification<br />

Classification of categories is the most widely used technique in medical data mining (Se-Chul Chun<br />

2008), and the C4.5 classification algorithms were used for the classification of diagnoses following<br />

feature extraction of medical data in order to make the classification. Results show this is consistent with<br />

the performances of these classification algorithms for other applications (Aurangzeb 2010).<br />

3.6.1 Rule base for probable diagnosis cardiovascular model<br />

To understand the rules for establishing the symptoms that are likely to cause cardiovascular disease, a<br />

method of constructing a rule-based decision tree from the medical terms was used to create a specific<br />

rule (Zhi-Hua Zhou et al. 2003). The rule was used to learn rules from the new training data set based on<br />

a technique of collecting “if . . . then . . .” rules (Kent W. Bridges 1992). The symptoms are variable,<br />

selected for the classification of cardiovascular disease and other diseases.<br />

406


Klaokanlaya Silachan and Panjai Tantasanawong<br />

3.6.2 Classification of document and diseases model<br />

In the present study we used two different methods for the classification of cardiovascular disease and<br />

medical document. They differ from traditional approaches by generating domain models from data. In<br />

the proposed system, we use two different machine learning methods for classifying: Decision Tree<br />

(C4.5) and Naïve Bayes. (J.R. Quinlan 1993; B.H. Jun et al.1997)<br />

C4.5 is a commercially available classification algorithm used to generate a decision tree using the idea<br />

of information entropy and was used for the analysis of a text categorisation system based on the<br />

decision tree model. The data set acquired from this study was used as an input to generate the decision<br />

tree using C4.5 algorithm. This research assumed uniform misclassification and performed a 10-fold<br />

validation for the classifier method.<br />

C4.5 defined the information gain formally as follows: (Anthony et al. 2004 and J.R Quinlan 1996).<br />

n<br />

Entropry(S) = Σ - Pi * log 2 Pi<br />

I=1<br />

Where: S = Data set, Sv = A partition of S according to the value of attribute A, n = The number of<br />

attributes A, |Sv| = The number of dataset in the partition, |S| = Total number of dataset in S.<br />

Gain(S,A) = Entropry(S)-Σ |S | Entropy(S )<br />

v v<br />

Value(A) |S|<br />

Where: X = Data set, n = The number of partition in S,pi = Proportion of Sv to S.<br />

Gain Ratio = Gain / Split Information<br />

Where split information is a value based on the column sums.<br />

A Bayes classifier is a simple probabilistic classifier based on Bayes theorem. The goal is to build a<br />

model. Given a set of data X = x1, x2, x3,…..,xn, a data mining problem is designed to uncover the<br />

properties of the distribution from which the set comes. The Bayes rule can be expressed as follows:<br />

P (h1|xi)= P (xi|h1)P (h1)/P (xi|x1)P (h1) + P (x1|h2)P (h2) (7)<br />

and called the posterior probability, while P(h1) is the prior probability. P(xi ) is the probability of the<br />

occurrence of data value and P(xi|h1) is the conditional probability ( Buntine W.1992 and Margaret H.<br />

2002).<br />

3.7 Performance evaluation<br />

Many different metrics are used in machine learning and data mining to build and evaluate models. The<br />

four performance measures including precision, recall, F-measure and accuracy are employed to<br />

evaluate models. A distinguished confusion matrix is obtained to calculate the four measures. The<br />

confusion matrix is a matrix representation of the classification results.<br />

Table1: A confusion matrix for a two-class classification<br />

Actual class<br />

Recognised<br />

Predicted as Positive Class Predicted as Negative Class<br />

Actual Positive class True Positive (TP) False Negative(FN)<br />

Actual Negative class False Positive(FP) True Negative(TN)<br />

Text classification rules are typically evaluated using performance measures from medical data. Common<br />

metrics for disease classification evaluation include recall, precision, accuracy F. Given a test set of N<br />

documents, the cells contain the counts for true positive (TP), false positive (FP), true negative (TN) and<br />

407<br />

(4)<br />

(5)<br />

(6)


Klaokanlaya Silachan and Panjai Tantasanawong<br />

false negative (FN), respectively. Clearly, N = TP + FP + TN + FN. The metrics for binary-decisions, the<br />

precision, recall, F-measure are defined as: (Gundersen et al.1996).<br />

4. Results<br />

TP<br />

Recall = (8)<br />

TN<br />

TP<br />

Precision = TP+FP<br />

(9)<br />

2*TP<br />

F-measure =<br />

2*TP+FP+FN<br />

(10)<br />

Accuracy = TN+TP (11)<br />

TN+ FN+FP+TP<br />

The medical dataset for our set of experiments is shown as follows:<br />

4.1 Extracted terms<br />

After extracting the symptom and disease terms, these can be reviewed as records and the<br />

corresponding terms as attributes of the record from the training dataset group. The example below<br />

shows the vocabulary used.<br />

Table 2: Sample of diagnosis single and multi-word<br />

weakness, paleness, chills, vomiting, cough, nausea, angina, syncope<br />

Single word<br />

Change in consciousness, shortness of breath, cardiac arrest, feeling of indigestion<br />

Multi-word<br />

Table 3: Samples of extracted symptom terms<br />

Documents Extracted Symptom Terms<br />

D-1<br />

chronic vapour, chest pain, low fever, panting, tired<br />

D-2<br />

headache, high fever, vapour, prickly heat<br />

D-3 encephalitis, have pains, pain and ache<br />

..... .......<br />

D-33 Angina, acute myocar(myocardium),backache, dizziness<br />

D-52 syncope, acute myocar<br />

Table 4: Sample document-term frequency data matrix<br />

Doc<br />

D-4<br />

D-33<br />

D-n<br />

TF1<br />

Angina<br />

TF2<br />

Syncope<br />

1 1<br />

...<br />

TFn<br />

Acute myocar<br />

1 - 1<br />

1 1 1<br />

408


4.2 Classification<br />

Klaokanlaya Silachan and Panjai Tantasanawong<br />

The implemented of C4.5 algorithm identified classifiers with weka tool from medical data. The medical<br />

data is divided into two sets of training set and test set. When the classification of diseases already will<br />

be able to see the pattern of the document and that the diagnosis of diseases The result of classification<br />

by disease in training set are shown as seen in Tables 5 and 6.<br />

Table 5: Sample of cardiovascular disease document classification<br />

Documents Cardiovascular Symptom Terms<br />

D-33<br />

D-52<br />

Angina, acute myocar, backache, Dizziness<br />

Syncope, Acute myocar<br />

D-87 Weakness, shortest of breath, angina<br />

Table 6: Sample of Cardiovascular disease classification<br />

Class Documents Symptom Terms<br />

Other<br />

Cardiovascular<br />

4.3 Evaluation performance<br />

D-1<br />

chronic vapour, chest pain, low fever, pant, be tired<br />

D-2 Have a headache, High fever, vapour, prickly heat, ..<br />

D-3 encephalitis , have pains, pain and ache, fever, itch<br />

........ ........<br />

D-33<br />

Angina, Acute myocar, backache, Dizziness<br />

D-52 Syncope, Acute myocar<br />

D-87 Weakness, shortest of breath, angina<br />

.. ........<br />

The data mining method used to build the model is classification. The instances in the dataset are<br />

representing the results of different types of testing to predict the accuracy of the cardiovascular disease<br />

document. The C4.5 and Naïve Bayes algorithms were used to predict the accuracy of cardiovascular<br />

medical dataset. This paper evaluation reports on the classifier are conducted with a Weka tool. The<br />

performance of the classifiers is evaluated and their results are analyzed. The results of the comparative<br />

analysis are based on 10 fold cross-validations. According to the attributes, the dataset divided into two<br />

parts such that 70% of the data is used for training and 30% is used for testing. For the purposes of this<br />

project, the data set was divided into a training set of 210 instances and a test set of 90 instances. See<br />

the performance comparison experiment in Table 7-8.<br />

Table 7: Evaluation on classification result using the training set<br />

C4.5 Naïve Bayes<br />

Precision 0.93 0.98<br />

Recall 0.95 0.95<br />

F-measure 0.97 0.97<br />

Accuracy 95.2 94.8<br />

Table 8: Evaluation of the accuracy of prediction results using the test set<br />

C4.5 Naïve Bayes<br />

Precision 0.96 0.95<br />

Recall 0.98 0.98<br />

F-measure 1 0.97<br />

Accuracy 99.8 99.6<br />

A comparison of the two algorithms with Precision, Recall and F-measure on medical training set and test<br />

set in a graph pattern seen in Figure 2 and 3.<br />

409


Klaokanlaya Silachan and Panjai Tantasanawong<br />

Figure 2: Precision, recall, f-measure of<br />

training set<br />

Figure 3: Precision, recall, F-measure of test<br />

set<br />

The results in Table 7-8 and Figure 2-3 show that the C4.5 algorithm can build the training set classifier<br />

better than Naive Bayes by a value 95.2%. In predicting the relationship of the symptoms to categories of<br />

cardiovascular disease, the test set has 99.8% prediction accuracy (percentage correct), and therefore is<br />

better than that of the Bayesian algorithm, with a precision rate of 0.96% and the recall at 1.0%. Finally,<br />

for the F-measure, which showed the relationship between the best queries and answer, recall is 0.98%.<br />

All results are acceptable.<br />

5. Conclusion<br />

This paper illustrates an approach where information extraction and features the selection of text medical<br />

data using sliding windows, TF-IDF techniques, with a multiword check and term weight-frequency.<br />

Information extraction(IE) is drawn from a medical text, and the data extraction is stored in a relational<br />

database of medical data. It is necessary to employ medical diagnostic data in the experimental<br />

evaluation. Data extraction is important for the classification of the documented information, including<br />

classification of whether or not cardiovascular disease is featured in major health information, called<br />

health text categories. It has been shown in this paper that classification, summary and probable<br />

diagnosis of cardiovascular disease can be carried out through data mining using rule induction. The<br />

commercially available C4.5 algorithm and Naïve Bayes algorithm were used as classifiers .The overall<br />

reliability was then checked and evaluated through a confusion matrix. precision, recall, F-measure, and<br />

accuracy were used as criteria for evaluating the performance of the training set with the accuracy values<br />

at 95.2%. The test set classifier for predictive accuracy of the C4.5 algorithm was marginally better than<br />

that of Bayes algorithm with the accuracy value at 99.8%. The follow-up to these test results would be to<br />

gain permission to evaluate the system against real data from a wide range of medical data records.<br />

References<br />

Anthony J. Myles, Robert N. Feudale, Yang Liu, Nathaniel A.Woody and Steven D. Brown. (2004) “An introduction to<br />

decision tree modeling”, [online] , Journal of chemometrics, Publish in Interscience.<br />

Appelt D, Israel D.(1999) ”Introduction to Information Extraction Technology”, In a tutorial for IJCAI-1999.<br />

Aurangzeb Khan, Baharum Baharudin, Lam Hong Lee, Khairrullah khan. (2010) “A Review of Machine Learning<br />

Algorithms for Text-Documents Classification”, Journal of Advance in information technology,VOL.1,No.1.<br />

B.H. Jun, C.S.Kim, H.Y.Song, J.Kim. (1997) “A new criterion in selection and discretization of attributes for the<br />

generation of decision tree”, IEEE Transaction on Pattern Analysis and Machine<br />

Intelligence,Vol.19,No.12,pp.1371-1375.<br />

Buntine W. (1992) “Learning Classification Trees”, Statistics and Computing.<br />

Dalila Bekhouche,Yann Pollet, Bruno Grilheres, and Xavier Denis. (2004) “Architecture of a Medical Information<br />

Extraction System”, NLDB2004, LNCS3136 Springer- Verlag Berlin Heidelberg, pp. 380-387, 2004.<br />

Dittapol Muntham, Lily Ingsrisawang. (2009) ”An Application of Decision Tree Algorithms for Diagnosis of the<br />

Respiratory System: A case Study of Pranakorn Sri Ayudthaya Hospital”, Journal of Health System, volume 1.<br />

Gundersen ML, Haug PJ, Pryor TA (1996) “Development and evaluation of a computerized admission diagnoses<br />

encoding system”, Comp Biomed Res, 29(5), 351–72.<br />

J.R Quinlan. (1996) ”Induction of Decision Tree”, Machine Learning, Vol, 1, No.1, pp.81-106.<br />

J.R Quinlan.(1993) ”C4.5: Programming for machine learning”, San Mateo,Calif,Morgan Kaufmann.<br />

Keku E, Rosamond W, Taylor HA Jr, Garrison R, Wyatt SB, Richard M, Jenkins B, Reeves L, Sarpong D. (2005)<br />

”Cardiovascular disease event classification in the Jackson Heart Study : methods and procedures”, Ethnicity &<br />

Disease, Volume 15, pp.73-76.<br />

Kent W. Bridges. (1992) “Rule-base Decision Making : A way to determine which alien species to control”, Alien Plant<br />

Invasions in Native Ecosystems of Hawaii: Management and Research.<br />

410


Klaokanlaya Silachan and Panjai Tantasanawong<br />

K.S. Kavitha, K.V.Ramakrishnan, Manoj Kumar Singh.(2010). ”Modeling and design of evolutionary neural network<br />

for heart disease detection”, IJCSI international journal of Computer science, Issue, Vol. 7, Issue 5.<br />

Margaret H., dunham. (2002) ”Data mining Introductory and Advanced Topics”, Chapter3, p.52.<br />

Maria Grineva, Maxim Grinev and Dmitry Lizorkin (2009) ”Effective Extraction of Thematically grouped Key Terms<br />

From Text “, Association for the Advancement of Artificial Intelligence.<br />

Parikshit Sondhi and Manish Gupta and ChengXiang Zhai and Julia Hockenmaier. (2010) “Shallow Information<br />

Extraction from Medical Forum Data”, Coling 2010: Poster Volume, pages 1158–1166, Beijing.<br />

Reddy,K,S. (2007) “India wakes up to the threat of cardiovascular diseases”. J.Am.Coll, Cardiol, 50, 1370-1372.<br />

S.Izumi, N.Yasuda, G.Itabashi, Y.Kato, K.Takahashi, T. Suganuma,N. Shiratori(2007) “A Health Advice Derivation<br />

System based on an Ontology”, 6 th International and 2nd Asian Semantic Web <strong>Conference</strong><br />

(ISWC2007+ASWC2007).<br />

Se-Chul Chun, Jin Kim, Ki-Baik Hahm, Yoon-Joo Park, Se-Hak Chun . .(2008) ”Data mining technique for medical<br />

informatics detecting gastric cancer using case-based reasoning and single nucleotide polymorphisms”, Expert<br />

Systems Volume 25, Issue 2, pages 163–172.<br />

T.Gruber (1995). “Toward Principles for the Design of Ontologies for Knowledge Sharing”, International Journal of<br />

Human - Computer studies, p.43, 907 – 928. WHO(1992). ” International Statistical Classification of Disease<br />

and Related Health Problem”, tenth revision (ICD-10), Geneve.<br />

Zhi-Hua Zhou, Yuan Jiang. (2003) “Medical diagnosis with C4.5 rule preceded by artificial neural network ensemble”,<br />

Information Technology in Biomedicine, IEEE Transactions on.Issue1. p.31-42.<br />

411


Optimizing Information Technology Value Governance<br />

Framework Based on Val IT<br />

Ali Suzangar, Mehrdad Kalantarian , Shirin Nasher , Mohammad Kajbaf and Negar<br />

Madani<br />

Infoamn IT Consultancy CO., Tehran, Iran<br />

a.suzangar@infoamn.com<br />

m.kalantarian@infoamn.com<br />

sh.nasher@infoamn.com<br />

m.kajbaf@infoamn.com<br />

n.madani@infoamn.com<br />

Abstract: Nowadays, information technology investment is crucial for organizations. Four IT investment principles<br />

are “risk”, “modeling”, “management” and “governance and compliance”. There are different frameworks such as<br />

ITIL, COBIT, Val IT and Risk IT which provide best practices to those IT investment principles. The Risk IT<br />

framework is about business risks related to the use of information technology. The connection to business is<br />

founded in the principles on which the framework is built, i.e. effective enterprise governance and management of IT<br />

risk. Val IT is a governance framework and complementary of COBIT from the business and financial perspective.<br />

Since Val IT framework only provides enterprise with the mechanism that it requires for monitoring and optimizing the<br />

IT business value, therefore there is no tool to calculate and evaluate information technology investment return. ITIL<br />

financial management process has three main aspects that provides guidance on how controlling spending and cost.<br />

In order to apply all direct and indirect cost evaluating in new information technology investment, a hierarchy decision<br />

making model is necessary. Since these frameworks activities are considered based on three principles of investing,<br />

i.e. risk, management and governance, none of them has a process or approach to create a model based on the<br />

investment costs in order to make final decision. So a comprehensive framework that based on all four investment<br />

principles, can be as a good solution for evaluating new information technology investments. In this paper, a RMMG<br />

framework that considers IT investment evaluation decision making model and financial process among the value<br />

governance and compliance activities, and also evaluating IT investment risk will be proposed to aid those<br />

organizations that confront by IT value investment issues.<br />

Keywords: IT value governance, Val IT, ITIL, Risk IT, decision making model, RMMG framework<br />

1. Introduction<br />

IT governance has been defined by ISACA as “the responsibility of executives and the board of directors,<br />

and consists of the leadership, organizational structures and processes that ensure that the enterprise’s<br />

IT sustains and extends the organizations strategies and objectives” (ISACA 2009a), (ITGI 2007b).<br />

IT governance can thus be pictured as focusing primarily on the following five areas (ISACA 2009a):<br />

Strategic alignment: achieving the goals and strategies of an enterprise through the coherent<br />

undertaking of activities by the different governance structures or management levels within an<br />

organization.<br />

Value delivery: creating new value for the enterprise through IT, maintaining and increasing value<br />

derived from existing IT investments, and eliminating IT initiatives and assets that are not creating<br />

sufficient value for the enterprise.<br />

Risk management: addressing IT-related risks. IT risk is the business risk associated with the use,<br />

ownership, operation, involvement, influence and adoption of IT within an enterprise.<br />

Resource management: ensuring that the right capabilities are in place to execute the strategic plan<br />

and sufficient, appropriate and effective resources are provided.<br />

Performance measurement: tracking the achievement of the objectives of the enterprise’s IT-related<br />

services and solutions and compliance with specific external requirements (ISACA 2009a).<br />

Three IT governance frameworks have been developed by ISACA: COBIT, Val IT and Risk IT. The<br />

integrated application of processes defined by COBIT, Val IT and Risk IT can help enterprises<br />

significantly improve IT governance and manage IT-related risks (ITGI 2007b).<br />

The best IT investments are those which help to maximize the value of the firm. They also contend that to<br />

maximize the value of the firm, IT investment decisions need to be able to maximize IT benefits while<br />

minimize IT risks (Joseph Wen, Yen and Lin 1998).<br />

412


Ali Suzangar et al.<br />

There are four main important issues about IT investment and return benefits as follows:<br />

1. Management of business risk is an essential component of the responsible administration of any<br />

enterprise. Almost every business decision requires the executive or manager to balance risk and<br />

reward (ISACA 2009c). The recognition of investment risk as an important component in IT<br />

investment decision making has long been recognized (Epstein and Rejc 2004-2005), (Joseph Wen,<br />

Yen and Lin 1998).<br />

2. Since organizations are often unable or unwilling to justify the expenditure to improve and develop<br />

services (OGC 2007), it is important to identify some methods and models to calculate the investment<br />

direct cost and indirect cost (tangible and intangible cost criteria).<br />

3. Always IT organizations looking for controlling the spending of money within the organization<br />

(annual budgeting), set of processes to account the budget that is allocated to information technology<br />

and also charging mechanisms to evaluate IT services delivered costs. As a result IT organizations<br />

need the financial management included budgeting, accounting and charging processes (OGC<br />

2007).<br />

4. IT organizations like to have governance framework for value management in a manner that is fully<br />

integrated with overall enterprise governance, provide strategic direction for the investment decisions,<br />

define the characteristics of portfolios required to support new investments and resulting IT services,<br />

assets and other resources and improve value management on a continual basis (ITGI 2008b).<br />

All above issues can be considered as four main IT investment principles as follows:<br />

1. Value risk assessment<br />

2. Value modeling<br />

3. Value management<br />

4. Value governance and compliance<br />

Since no existence value frameworks can cover all mentioned principles, this paper will tend to propose a<br />

comprehensive framework based on these principles.<br />

In literature review, first Risk IT framework will be defined. Then, an analytic hierarchy IT investment<br />

decision making model based on multi criteria multi objective method will be introduced. Note that since<br />

the designing methodology of an analytic hierarchy IT investment decision making model is not the main<br />

concern of this paper, it would be introduced briefly in 2.2. But the complete approach has been<br />

submitted as a separate paper to this conference titled “Evaluation of IT investment methods and<br />

proposing a decision making model”. After that, ITIL financial management process will be defined in<br />

three parts to control spending costs in the organization. In the consequence, Val IT framework will be<br />

described as a mechanism for many organizations to assess their current IT value state and recognize<br />

their pain points, and other symptoms in creating IT enabled value.<br />

After comparing these frameworks, the RMMG proposed framework will be defined with its related layers<br />

and processes. In conclusion, the advantages of proposed RMMG will be described.<br />

2. Literature review<br />

2.1 Risk IT framework<br />

The all-encompassing use of IT can provide significant benefits to an enterprise, but it also involves risk.<br />

Due to IT’s importance to the overall business, IT risk should be treated like other key business risks,<br />

such as strategic risk, environmental risk, market risk, credit risk, operational risks and compliance risk,<br />

all of which fall under the highest ‘umbrella’ risk category. While these other risks have long been<br />

incorporated into corporate decision-making processes, too many executives tend to relegate IT risk to<br />

technical specialists outside the boardroom (ISACA 2009c).<br />

The Risk IT framework is about IT risk, in other words, business risk related to the use of IT. The<br />

connection to business is founded in the principles on which the framework is built, i.e., effective<br />

enterprise governance and management of IT risk. The three domains of the Risk IT framework are as<br />

follows: Risk Governance, Risk Evaluation and Risk Response (ISACA 2009c).<br />

413


Ali Suzangar et al.<br />

The risk management process model groups key activities into a number of processes. These processes<br />

are grouped into three domains. The process model will appear familiar to users of COBIT and Val IT<br />

(ISACA 2009c).<br />

2.2 Analytic hierarchy it investment decision making model<br />

Since the intangible criteria are complex to evaluate, therefore this action may take a long time and<br />

efforts. So many managers prefer to neglect them and just focus on the financial factors (Borenstein<br />

2005). They think that IT investment affects on their organization and enhances productivity efficiency in<br />

situ. But indeed return on IT investment has a potential latency to create value just because of these<br />

intangible factors. So it shows the importance of these criteria to have an appropriate investment<br />

evaluation. The existence of the pre-defined criteria hierarchy model based on the consensus of many<br />

specialist, can structured the decision making process systematically. Figure1 shows this proposed<br />

model.<br />

For this model, first tangible and intangible criteria were identified and then categorized in different criteria<br />

and sub criteria. As a consequence, an importance rank was determined to them according to the<br />

experiment of chief managers. Based on these ranking, at last this hierarchy model in order to consider<br />

all these tangible and intangible criteria in IT investment evaluation was proposed. Note that this<br />

proposed model is general model and independent from the specific IT investment to aid investment<br />

decision making process.<br />

Figure 1: Analytic hierarchy decision making model<br />

So there was a consensus that this proposed model has improved the decision process in order to have<br />

consistent and complete criteria based on different tangible and intangible factors for IT investment<br />

evaluation.<br />

414


2.3 ITIL financial management process<br />

Ali Suzangar et al.<br />

Financial Management is the sound stewardship of the monetary Resources of the organization. It<br />

supports the organization in planning and executing its business objectives and requires consistent<br />

application throughout the organization to achieve maximum efficiency and minimum conflict.<br />

Within an IT organization it is visible in three main processes as follows: Budgeting, IT accounting, and<br />

charging (OGC 2007).<br />

2.3.1 Budgeting<br />

Is the Process of predicting and controlling the spending of money within the organization and consists of<br />

a periodic negotiation cycle to set budgets (usually annual) and the day-to-day monitoring of the current<br />

budgets.<br />

The aim of Budgeting is that the actual costs match the budget (predicted costs). This budget is usually<br />

set by negotiations with the Customers who are providing the funds. Good Budgeting is essential to<br />

ensure that the money does not run out before the period end. Where shortfalls are likely to occur the<br />

organization needs early warning and accurate information to enable good decisions to best manage the<br />

situation (OGC 2007).<br />

2.3.2 IT accounting<br />

IT Accounting is the set of processes that enable the IT organization to account fully for the way its<br />

money is spent (particularly the ability to identify costs by Customer, by service, by activity). It usually<br />

involves ledgers and should be overseen by someone trained in accountancy (OGC 2007).<br />

2.3.3 Charging<br />

Charging is the set of processes required to bill Customers for the services supplied to them. To achieve<br />

this requires sound IT Accounting, to a level of detail determined by the requirements of the analysis,<br />

billing and reporting processes (OGC 2007).<br />

2.4 Val IT<br />

Executives, even if they are aware of the need for more effective governance and management of<br />

information technology, may not recognize that many of the day-to-day business challenges they face<br />

involve issues of value management. Val IT provides proven value management principles, processes<br />

and practices to enable enterprises to maximize the delivery of business value from investments involving<br />

IT (ISACA 2010), (ITGI 2008c).<br />

There are some internal trigger (major IT project failures, serious budget overruns and etc) and external<br />

events (a major shift in the marketplace and etc) that are likely to convince executives to improve their<br />

enterprise value management practices (ISACA 2009b).<br />

After identifying these challenges within organization, the Val IT framework provides useful guidance on<br />

proven processes and practices that enable effective governance of investments involving IT.<br />

This guidance is found in the three domains of Val IT (ITGI 2008b):<br />

Value Governance (VG): ensuring that value management practices are embedded in the enterprise<br />

Portfolio Management (PM): ensuring that the enterprise secures optimal value across its portfolio of<br />

investments involving IT<br />

Investment Management (IM): ensuring that the enterprise’s individual investments each contribute<br />

the value expected of them.<br />

Val IT Processes are a collection of interacting activities undertaken in accordance with management<br />

practices (ITGI 2008b).<br />

3. Comparing<br />

According to previous discussion, it can be considered four main principles in IT investment discussing,<br />

i.e. “value risk assessment”, “value modeling”, “value management” and “value governance and<br />

compliance”. Each mentioned frameworks focuses on one or more than one of these principles. So in<br />

415


Ali Suzangar et al.<br />

Table 1, these frameworks are compared in order to show how much they cover the aspects of<br />

investment evaluation. The main target of the frameworks and also the processes and domains goals that<br />

are included in those frameworks are assumed as this table comparison criterion.<br />

Table1: comparison table<br />

Val IT<br />

ITIL<br />

Risk IT<br />

Value Risk Assessment<br />

Low<br />

Low<br />

High<br />

Value Modeling<br />

Non<br />

Non<br />

Non<br />

Value Management<br />

Medium<br />

High<br />

Low<br />

Value Governance and compliance<br />

High<br />

Medium<br />

Low<br />

Based on this table, all frameworks have concentrate on each specific principle. It means that e.g. Val IT<br />

focus on the governance aspect is high, since management and risk aspect is not considered in the high<br />

level. The main point of this table is that none of them concentrate on the value modeling principle.<br />

4. Proposing RMMG framework<br />

The proposed framework scenario is defined as follows:<br />

Assume that, a new IT project is proposed to an organization to invest on it or the organization decides to<br />

invest on a new IT project based on its own objectives and strategies. By this trigger, some readiness<br />

activities should be done providing the safe investment and appropriate evaluation for managers. As an<br />

important principle, organization requires assessing risks before and after new investment to monitor and<br />

evaluate all investment opportunities and threats in order to able mitigating and responding them. Then<br />

beside the risk assessment, a model based on tangible and intangible benefit factors of proposed IT<br />

project can help managers to make a final decision on allocating their budget or not.<br />

Additionally, the organization requires a management process that can predict and control its costs and<br />

also account fully for the way its money is spent. So this process in order to provide requirement of IT<br />

investment justification is demanded. And at last, value governance and compliance is applied to take<br />

full advantage of proposed IT project, maximizing benefits, capitalizing on opportunities and gaining<br />

competitive advantages.<br />

By this scenario, the organization will require an overall comprehensive framework that is consisted of<br />

value risk assessment, value modeling, value management and value governance and compliance.<br />

Figure 2 illustrates this proposed framework that is called Risk, Modeling, Management and Governance<br />

(RMMG).<br />

The RMMG consists of four layers as Figure 2 illustrates. Each provides the guidance necessary for a<br />

comprehensive framework as required by the mentioned investment principles:<br />

Value Risk Assessment<br />

Value Modeling<br />

Value Management<br />

Value Governance and Compliance<br />

RMMG framework structure has been fully changed against Risk IT, ITIL, and Val IT frameworks. But<br />

those framework processes can be located in different layers of the RMMG as follows based on the main<br />

goal of the processes, and also the domains targets that are included in those frameworks.<br />

Each layer addresses capabilities having direct impact on the new information technology project<br />

investments. The structure of the layer is in the form of a lifecycle. It is iterative and multidimensional. The<br />

layer is expected to provide structure, stability and strength to investment capabilities with durable<br />

principles, methods and tools. This serves to protect investments and provide the necessary basis for<br />

value risk, management and governance. The RMMG can be adapted for use in various business<br />

environments and organizational strategies. The layers of RMMG definitions are as follows:<br />

416


Figure2: RMMG framework<br />

4.1 Value Risk Assessment<br />

Ali Suzangar et al.<br />

The Value Risk Assessment layer, as the core of the framework, provides how assessing investment<br />

risks before and after investing and implements activities to analyze and react to those risks. Value Risk<br />

Assessment is useful in the context of Value Modeling, Value Management, and Value Governance and<br />

Compliance. Value Risk Assessment defined processes that are as the same as Risk IT framework<br />

related processes as follows:<br />

Establish and maintain a common risk view (Risk IT)<br />

Integrate with ERM (Risk IT)<br />

Make risk-aware business decisions (Risk IT)<br />

Collect data (Risk IT)<br />

Analyze risk (Risk IT)<br />

Maintain risk profile (Risk IT)<br />

Articulate risk (Risk IT)<br />

Manage risk (Risk IT)<br />

React to events (Risk IT)<br />

4.2 Value Modeling<br />

The Value Modeling layer is considered in two parts, i.e. direct evaluation and indirect evaluation in order<br />

to offer the analytic hierarchy decision making model based on these two evaluation types of investment.<br />

In the indirect evaluation part, it covers all existence investment intangible criteria for converting them to<br />

quantitative cost. The scope of this layer includes developing a decision making model for managers to<br />

understand the weight of each investment criteria according to their objectives and strategies. The<br />

complete approach has been submitted as a separate paper to this conference titled “Evaluation of IT<br />

investment methods and proposing a decision making model”.<br />

417


4.3 Value management<br />

Ali Suzangar et al.<br />

The value management layer practices includes controlling the spending of money within the organization<br />

(annual budgeting), set of processes to account the budget that is allocated to information technology<br />

and also charging mechanisms to evaluate IT services delivered costs. It provides the strategic direction<br />

for investments, the desired characteristics of the investment portfolio, and the resources and funding<br />

constraints within which portfolio decisions must be made. Also, defines potential programs based on<br />

business requirements, determines whether they are worthy of further consideration, and develops and<br />

passes business cases for candidate investment programs to portfolio management for evaluation. One<br />

of the processes of this layer is from ITIL and the others are Val IT related processes as follows:<br />

Financial Management (ITIL)<br />

Establish strategic direction and target investment mix (Val IT)<br />

Determine the availability and sources of funds (Val IT)<br />

Manage the availability of human resources (Val IT)<br />

Evaluate and select programs to fund (Val IT)<br />

Monitor and report on investment portfolio performance (Val IT)<br />

Optimize investment portfolio performance (Val IT)<br />

Develop and evaluate the initial program concept business case (Val IT)<br />

Understand the candidate program and implementation options (Val IT)<br />

Develop the program plan (Val IT)<br />

Develop full life-cycle costs and benefits (Val IT)<br />

Develop the detailed candidate program business case (Val IT)<br />

Launch and manage the program (Val IT)<br />

Update operational IT portfolios (Val IT)<br />

Update the business case (Val IT)<br />

Monitor and report on the program (Val IT)<br />

Retire the program (Val IT)<br />

4.4 Value Governance and Compliance<br />

The Value Governance and Compliance layer establishes the overall framework, including defining the<br />

portfolios required to manage investments and resulting IT services, assets, and resources. This layer<br />

monitors the effectiveness of the overall governance framework and supporting processes, and<br />

recommends improvements as appropriate. The processes of this layer are as same as Val IT value<br />

governance related processes as follows:<br />

Establish informed and committed leadership (Val IT)<br />

Define and implement processes (Val IT)<br />

Define portfolio characteristics (Val IT)<br />

Align and integrate value management with enterprise financial planning (Val IT)<br />

Establish effective governance monitoring (Val IT)<br />

Continuously improve value management practices (Val IT)<br />

5. Conclusion<br />

Organizations apply different IT governance and management frameworks in order to reach their<br />

objectives and strategies such as improving their services and gaining competitive advantages, through<br />

new information technology investment decisions. According to this paper, we can consider four main<br />

principles to evaluate a new IT project investment, i.e. Value Risk Assessment, Value Modeling, Value<br />

Management and Value Governance and Compliance. There are different frameworks based on these<br />

principles. Since each of these frameworks considers one or two mentioned evaluating principles, so the<br />

existence of the comprehensive framework can be applicable for those organizations that want to<br />

implement an appropriate processes and activities to have an appropriate and exact evaluating in all new<br />

418


Ali Suzangar et al.<br />

investment decisions. Hence in this paper, a new conceptual and comprehensive framework was<br />

identified and illustrated by four layers in order to cover and consider all those principles that was called<br />

RMMG framework.<br />

This proposed framework, in addition to assess value risk, implement value management and<br />

governance activities, it evaluates direct and indirect investment costs to create an analytic hierarchy<br />

decision making model to help organization in their investments. It states the RMMG strength point.<br />

References<br />

Borenstein, D. and Baptista Betencourt, P. R. (2005) “A Multi-criteria Model for the Justification of IT Investments”,<br />

INFOR Journal, Vol. 43, No. 1, Feb, pp. 1-21.<br />

Epstein, M. and Rejc, A. (2004-2005) “Measuring the payoffs of IT investments”, CMA Management Journal, Vol. 78,<br />

No. 8, pp. 20-25.<br />

ISACA (2009a) Implementing and Continually Improving IT Governance, Rolling Meadows, IL 60008 USA.<br />

ISACA (2009b) IT Value Special Compilation, Rolling Meadows, IL 60008 USA.<br />

ISACA (2009c) The Risk IT Framework, Rolling Meadows, IL 60008 USA.<br />

ISACA (2010) Value Management Guidance for Assurance Professionals: Using Val IT 2.0, Rolling Meadows, IL<br />

60008 USA.<br />

ITGI (2007b) IT Governance Roundtable: IT Governance Trends, Rolling Meadows, IL 60008 USA.<br />

ITGI (2008a) Enterprise Value: Governance of IT Investments Getting Started With Value Management, Rolling<br />

Meadows, IL 60008 USA<br />

ITGI (2008b) Enterprise Value: Governance of IT Investments the Val IT Framework 2.0, Rolling Meadows, IL 60008<br />

USA<br />

ITGI (2008c) IT Governance Global Status Report, Rolling Meadows, IL 60008 USA.<br />

Joseph Wen, H., Yen, D.C. and Lin, B. (1998) “Methods for measuring information technology investment payoff”,<br />

Human Systems Management Journal, Vol. 17, No. 2, January, pp. 145–153.<br />

OGC (2007) Service Design ITIL service management practices v.3, London: the stationary office.<br />

419


Developing an Outpatient Electronic Medical Record System<br />

in Taiwan From a Physicians’ Perspective<br />

Hsiao-Ting Tseng, Pei-Ting Chang and Ray Chang<br />

National Chung Cheng University, Chia-yi, Taiwan<br />

appleapple928@gmail.com<br />

chang.payting@gmail.com<br />

misicc.pig@gmail.com<br />

Abstract: Electronic medical record systems (EMRS) are regarded as an innovative administrative tool in healthcare<br />

to integrate patient information. Physicians can use an EMRS to receive integrated information to provide better<br />

outpatient service. Many hospitals worldwide, including in Taiwan, have either implemented or plan to implement<br />

EMRS to reduce or replace the traditional paper-based charts and patient records. The key reason for some failures in<br />

implementation of EMRS is that system design was not user-friendly, i.e. it did not take into account physicians’ real<br />

needs. This research employed a case study to explore the development of an Outpatient Electronic Medical Record<br />

System (OEMRS) that was developed on the basis of physicians’ expectations. A qualitative interview confirmed that<br />

physicians’ participation can significantly affect their satisfaction and acceptance toward OEMRS use.<br />

Keywords: physician’s perspective, electronic medical record system, outpatient, satisfaction, acceptance<br />

1. Introduction<br />

According to the annual report of the Institute of Medicine for the year 1999, an estimated 98,000 people<br />

died because of medical negligence directly or indirectly. With IT applications, the casualty figure can be<br />

reduced to 55% (Drew, Carolyn and Robert, 2004). IT applications can improve the quality of medical care<br />

significantly. Since then, application of information technology to improve healthcare efficiency and to<br />

reduce healthcare expenditure has become an important trend among healthcare institutions.<br />

Electronic medical records systems are increasingly being used to keep record of medical procedures,<br />

interactions, treatment and procedures by physicians, medical staff and even patients. Using<br />

computer-based technologies to convert traditional medical records into electronic file formats, and to<br />

allow medical personnel to access this system via computers is called electronic medical records or EMR.<br />

According to the Medical Records Institute, electronic medical records are defined as electronic<br />

information of an individual’s lifelong health records. The coverage of Electronic Medical Records (EMR)<br />

includes automatic formatting of paper-based records and other health information, such as medication<br />

records, examinations charts and images of X-rays, etc.<br />

EMR have been used at all-levels in healthcare institutions to cope with a variety of medical service<br />

demands, such as teaching, evaluation, telemedicine and chronic care. According to a survey by the<br />

Bureau of National Health Insurance in 1999, about 98% of Taiwan’s medical institutions have introduced<br />

at least one (of several) computer-based technology to assist medical practitioners. EMR can reduce<br />

operational costs in hospitals and save space required for storage of physical paper-based records;<br />

physical space availability is very limited in Taiwan. Furthermore, patients’ medical records in different<br />

healthcare institutions can be integrated by EMR. It can not only reduce wasteful consumption of medical<br />

resources but also provide patients with continuous and high-quality medical services.<br />

The main users of EMR are healthcare staff and physicians. Active participation of physicians in the<br />

process of promoting electronic medical records can significantly enhance application and introduction of<br />

EMR in the healthcare industry (citation). Peter et al. (2006) showed that active participation of<br />

physicians in the process of introduction of EMR can help make the interface of EMR systems more<br />

suitable for physicians’ needs. Moreover, with participation of physicians, system implementation can<br />

become a much smoother process. It is important for a EMR system to meet physicians’ needs and yet<br />

there are not many studies on doctors’ views on efficacy of EMR systems. In this study, we examine the<br />

working of a hospital which has developed an Outpatient Electronic Medical Records System (OEMRS) as<br />

a case study to explore and understand the development and evaluation of the system from physicians’<br />

perspectives.<br />

2. Background<br />

Electronic medical records include records of all healthcare related information required by clinical workers<br />

for providing effective and continuous healthcare. Electronic medical records not only constitute an<br />

420


Hsiao-Ting Tseng et al.<br />

electronic record of diagnoses and prescribed treatments, etc., but they also facilitate medical information<br />

integration. The goals of this system are storage and management of medical history of individual<br />

patients, as well as medical knowledge used for their treatments at different times. Hence, EMR can help<br />

healthcare personnel make related decisions and improve the quality of medical care, besides making<br />

hospital management more efficient. Electronic medical records should focus on one objective: to record<br />

long-term medical care information contiguously to provide continuous, useful and high-quality medical<br />

care.<br />

In the system’s development phase, participation of users is regarded as the key factor for success. Lee<br />

and Kim (1999) showed that good communication between information workers and users can have<br />

positive effect on eliminating perception differences (Lee and Kim, 1999). The literature also indicates that<br />

in the process of system development, the more there is users’ participation, the deeper is the influence<br />

on the final product (Zmud and Cox, 1979).<br />

In an outpatient medical care system, physicians are responsible for providing medical care services like<br />

prescriptions and examinations. Prior studies have showed that if physicians participate in the<br />

development process of the OEMRS, the interface of OEMRS can meet the physicians’ needs, and thus<br />

the need for educating and training them to use the OEMRS can be reduced (Peter et al., 2006). Using an<br />

electronic medical record system, physicians’ work efficiency, data processing procedures, and time<br />

management can be greatly improved (Richard, Elaine and Don, 1997). However, the most important<br />

concerns of physicians about OEMRS are patient safety and the working process. These factors can<br />

result in impedance to introduction of an OEMRS (Robert, 2006). Thus, if we physicians act as advocates<br />

or consultants for an electronic medical system, resistance can be reduced (Peter et al., 2006).<br />

3. Material and methods<br />

Most previous research about electronic medical records has focused on technical perspectives,<br />

technology combination, sharing mechanism, related standards and legal perspectives such as privacy<br />

concerns. Few studies have tried to understand the role physicians’ participation can play in the<br />

development process. Past research has showed that lower participation of medical staff can be<br />

considered a dilemma while planning introduction of electronic medical records, whether in small clinics or<br />

large hospitals. Therefore, this research has chosen to examine the case of a hospital that invited<br />

physician participation in developing an OEMRS.<br />

3.1 Case hospital<br />

The subject 1,000-bed hospital is located in southern Taiwan. The outpatient department of the hospital<br />

serves approximately 3,600 people daily and offers over 40 medical specialties. A new building was<br />

recently constructed to offer more outpatient services and preventive medical treatment.<br />

Both computerized and paper-based medical records are used in the outpatient department of the<br />

hospital. While a computerized physician order entry system is in use, paper-based medical records can<br />

still be obtained for use by physicians as references for patients’ histories. Outpatient consultation,<br />

emergencies, admission records, examinations images and pathology reports are available in physical<br />

form. Past records cannot be displayed on the same page when using the same program for browsing. If<br />

physicians need to query some related information, they need to access a sub-system or alternate<br />

windows. Due to the inconvenience of this process, most physicians still use paper medical records. In<br />

response to expansion of outpatient department services in the future, and to address difficulties in<br />

cross-referencing of medical records, it is necessary to develop an integrated outpatient electric medical<br />

records system.<br />

A high compliance rate was observed among information department staff and clinic physicians at the<br />

case hospital. The medical vice-president and the administrative vice-president showed great interest in<br />

this project and provided all support, hoping that the OEMRS can solve the difficulties of cross-referencing<br />

of medical records. A project team was formed with the deputy superintendent of the hospital as the<br />

convener for implementation of the OEMRS at the case hospital. Representatives from the administrative<br />

department were the president’s special assistant and five members from the information department.<br />

One physician was chosen from each major department as the consultant physician. All consultant<br />

physicians were active and showed strong interest in OEMRS development. In addition, they interacted<br />

well with the information department staff. Consultant physicians said they would like to participate in the<br />

421


Hsiao-Ting Tseng et al.<br />

development process of the OEMRS to further solve the difficulties of cross-referencing of medical<br />

records.<br />

3.2 System development<br />

In the process of development of OEMRS, the clinical processes that physicians wanted covered were<br />

listed and incorporated in the OEMRS to make the clinical process smooth. Secondly, in-depth interviews<br />

and questionnaires were used for examining the outpatient department’s work flow and practices. System<br />

development steps are shown in Figure. 1.<br />

Figure 1: System development steps<br />

The prototyping life-cycle, brought by Boar (1984) was employed as a system development method for<br />

this study. A visible and operable prototype was used as a communication tool between developers and<br />

related users (e.g. physicians). As a result, the time and cost of system analysis can be reduced. There<br />

are 3 factors affecting the decision whether an information system is suitable for the prototyping method<br />

(Table 1): (1) degree of environmental change, (2) degree of structuring of the decision-making process,<br />

and (3) similar previous design experiences (Kendall and Kendall, 1988).<br />

Table 1: Factors of system usage<br />

422


Hsiao-Ting Tseng et al.<br />

The main reasons for choosing the prototyping method to develop the OEMRS are as follows.<br />

Demands from physicians are often not clear, or are vague, and these demands cannot be listed easily.<br />

Meanwhile, users’ demands may also change over time. In addition, outpatient department’s physicians in<br />

the case hospital have long experience in clinical work but lack experience in OEMRS development and<br />

usage. Required features and specifications of an electronic medical records system may not be easily<br />

identified. Using the prototyping life cycle system, the system can be modified repeatedly to fit users’<br />

demands. Also, suitability of system interface can be confirmed with users to obtain better suggestions for<br />

further modifications. Based on a deep study of working of the case hospital’s outpatient department and<br />

interviews with related users, the users’ demands were identified as shown in Table 2.<br />

Table 2: System demands<br />

The needed software/hardware are tabulated in Table 3.<br />

Table 3: System develop environment and tools<br />

423


Hsiao-Ting Tseng et al.<br />

The OEMRS integrates other systems as shown in Figure 2. After patient’s registration, his/her basic<br />

information is recorded (or retrieved from existing records). If physiological or biochemical or radiological<br />

examinations are required, examination sheets are signed by the concerned department. Data are then<br />

transmitted back to the OEMRS through PACS. Physicians can see all records of patients’ diagnoses and<br />

determining treatments.<br />

Figure 2: OEMRS process diagram<br />

The main functions of OEMRS are shown in Figure 3. They are: (1) Logon portal; (2) Medical record main<br />

page; (3) History of visit records; (4) Special medication records (offering special chemical treatment<br />

medicine record and general medicine query function); (5) Examination and check reports; (6) Case<br />

management records (offering case management records including cancer cases, chronic kidney disease<br />

and diabetes cases); (7) Special check records (offering special check reports).<br />

Figure 3: OEMRS function infrastructure diagram<br />

424


Hsiao-Ting Tseng et al.<br />

The OEMRS integrates the above functions in a window on the computer screen which can be adjusted to<br />

meet the physician’s needs. Major treatment records and related information for a patient are put together<br />

on a first screen with specific font colors or alert windows to remind physicians of patient previous<br />

treatment histories.<br />

Physicians can log on to OEMRS from the original computerized physician order entry (CPOE) or PACS.<br />

After logging on, users can input patients’ medical record numbers to find related information and view<br />

PACS images. If data are input into the OEMRS via the original interface, medical record numbers of<br />

patients are directly used by the OEMRS to show the related information. Discharge medical note shows<br />

information about a patient’s diagnostic records and examinations from admission to discharge, disease<br />

history, and discharge medication records. Operations notes provide patients’ surgery related records to<br />

meet surgeons’ needs. As a result, timely viability of surgery records and operations notes queries can be<br />

fulfilled.<br />

Outpatient physicians can browse patients’ history to check outcomes of previous visits and compare<br />

different records to observe changes of chief complaints in order to adjust medication. Special<br />

medication records answer any queries for special chemical treatments which are distinguished from<br />

general medicine. Furthermore, it can be sorted and arranged by medicine name, department, or date of<br />

prescription to meet physicians’ needs. Outpatient physicians can quickly read check reports through<br />

Examination by choosing specific data bases to browse information they need. Pre-established<br />

sequencing patterns are listed by date, or by users needs. Extreme Values higher than standard are<br />

shown as red (H), and those of lower values are shown as red (L) to alert physicians. This function can<br />

fulfill physicians’ needs for collection of related continuous examination data, neurological examination<br />

data, and hints of extraordinary examination results. Physicians usually have to compare different<br />

examination reports of different dates in the outpatient consulting process for determining further<br />

treatment. Reports of the same items of examination and tests carried out at different times are compiled<br />

in the same window. Special items are shown in red fonts to alert the physicians. Examination statistics<br />

can be converted into line charts by physicians to help patients understand the changes in their<br />

examination data. Such graphical presentations of examination reports help pediatric doctors explain<br />

patients’ condition to their parents. Case management records and collects data that a case manager<br />

needs about patient-related information such as cancer case management, chronic kidney disease, and<br />

diabetes case management. Special examination records can be searched from the main page tag.<br />

4. System evaluation<br />

Since physicians usually have higher control over medical systems, and are always critical in deciding<br />

whether the new system can be promoted smoothly or not, the case hospital developed the OEMRS from<br />

a physicians’ perspective. During system planning and development phase, modifications are made after<br />

discussions on the OEMRS between hospital personnel from different departments and the system<br />

vendor. These modifications and discussions were analyzed, developed and practiced by the information<br />

staff. After the system went online, factors influencing system usage and feedback from users were taken<br />

for future improvements to realize the system’s benefits. Hence, in-depth interviews were used in this<br />

research to get individual opinions that the questionnaire cannot get. This research used semi-structured<br />

interviews and selected pediatric physicians as interviewees since that was the most supportive division in<br />

the outpatient department of the subject hospital. Usages of OEMRS by the interviewers are shown in<br />

Table 4.<br />

Most interviewees started to use the OEMRS in 2008, and were using the system actively at the time of<br />

the interview. The reason to use OEMRS was the hospital’s policy stance and the consulting physician<br />

having joined the OEMRS implementation team. Except the director of the division, none of the physicians<br />

tried to influence others to use the system. Table five shows individual evaluations of the OEMRS.<br />

Table five shows that the quality of the process of information storage and retrieval by the OEMRS was<br />

still below the physicians’ expectations. Most interviewees reported “partial” in stead of positive to the<br />

OEMRS information quality. System quality of OEMRS was assured by most of the interviewees which<br />

contributed to continuous use of the OEMRS.<br />

425


Table 4: Usage of OEMRS by interviewees<br />

Table 5: Evaluation of OEMRS<br />

Hsiao-Ting Tseng et al.<br />

426


5. Conclusion and discussion<br />

Hsiao-Ting Tseng et al.<br />

The objective of this research is to use a case study to explore the implementation of an OEMRS with<br />

physician’s participation. The interviews show that one of the motivations for physicians to use OEMRS is<br />

to accept and promote hospital policy. In the development phase of the OEMRS, cooperation and support<br />

of consultant physicians of the case hospital and communication between physicians and information staff<br />

were critical factors for the success of system development and operation. The system development<br />

experiences can serve as references for other hospitals in developing their own OEMRS.<br />

It is clear that physicians’ satisfaction will be improved if the system interface is easy to use and<br />

corresponds to physician needs. Usage frequency of OEMRS is directly influenced by user experience.<br />

More improvements and modifications need to be done to increase physicians’ acceptance and OEMRS<br />

usage. In summary, physicians’ attitudes toward OEMRS development are generally positive. How to<br />

upgrade OEMRS to assist more physicians’ clinical decision making is the task to be considered. For<br />

instance, integrating automated health examination systems (Kuo and Fuh, 2010) to enhance inference<br />

quality and wireless/mobile devices to benefit point of care (Chung, Bell and Lee, 2006) are feasible<br />

directions.<br />

References<br />

Boar B. (1984) Application Prototyping: A Requirements Definition Strategy for the 80s, John Wiley & Sons, USA, New<br />

York.<br />

Chung, K., Bell, R. and Lee, D. (2006) “A Point of Care Clinical Documentation System for Hospice Care Providers”,<br />

Journal of Medical Systems, Vol. 30, No. 1, pp 33-37.<br />

Drew E. A., Carolyn C. and Robert J. B. (2004) “Improving Patient Safety — Five Years after the IOM Report.” The<br />

New England Journal of Medicine, Vol. 35, No. 1, pp 2041-2043.<br />

Lee, J. N. and Kim, Y. G. (1999) “Effect of Partnership Quality on IS Outsourcing: Conceptual Framework and<br />

Empirical Validation.” Journal of Management Information Systems, Vol. 15, No. 4 pp 29-62.<br />

Kendall, E. K. and Kendall, E.J. (1988) Systems Analysis and Design, Englewood Cliffs, Prentice-Hall, Boston.<br />

Kuo, K.L. and Fuh, C.S. (2010) “A Health Examination System Integrated with Clinical Decision Support System.”<br />

Journal of Medical Systems, Vol. 34, No. 5, pp 829-842.<br />

Peter G. R, Howard A., Robert W. and Peter B. (2006) “Active Physician Participation Key to Smooth MOE MAR<br />

Rollout.” Healthcare Quality, Vol. 10, sp, pp 58-64.<br />

Richard S. D., Elaine B. S. and Don E. D. (1997) The Computer-based Patient Record: An Essential Technology for<br />

Health Care. National Academy Press, Washington.<br />

Robert H. M. and Ida S. (2004) “Physicians' Use of Electronic Medical Records Barriers and Solutions.” Health Affairs,<br />

Vol. 23, No. 2, pp 116-126.<br />

Robert M. W. (2006) “Expected and Unanticipated Consequences of the Quality and Information Technology<br />

Revolutions.” Journal of the American Medical Association, Vol. 296, No. 20, pp 2780-2783.<br />

Rogers, E. M. (1995) Diffusion of Innovations (4th ed.), Free Press, New York.<br />

Zmud, R. W. and Cox, J. F. (1979) “The implementation process: A change approach”. MIS Quarterly, Vol. 3, No. 2,<br />

pp 35-43.<br />

427


MTN Foundation's Digital Library Project in Nigerian<br />

Universities: An Evaluative Study<br />

Ngozi Blessing Ukachi<br />

University of Lagos, Akoka, Nigeria<br />

ukachingozi2001@yahoo.com<br />

Abstract: This survey study was conducted with the main objective of evaluating the use and service provision of<br />

MTN Foundation’s digital library project in Nigerian universities with regards to the objectives of establishing them.<br />

The initial plan to study two out of the three already commissioned libraries was forestalled by temporary closure of<br />

the second one hence the study concentrated on one, situated at University of Lagos, Nigeria. Questionnaire<br />

constituted the major instrument for data collection even though the librarian in charge of this library was interviewed<br />

and documentary records consulted to obtain such data as list of electronic resources/database available in the<br />

library. A total number of 1000 questionnaires were distributed to the library users. Out of this number, 852 were<br />

returned while 812 questionnaires were found usable. The data collected was subjected to descriptive statistical<br />

analysis using simple percentages and frequency counts. The result of the study revealed that this project is<br />

appreciated by users considering their level of acceptance of the fact that; the resources in this library has enriched<br />

their awareness on the existence of various kinds of e- resources (76.6%); the use of the resources has assisted<br />

them greatly in doing their assignments (68.8%); and resources from this library have made their research work very<br />

easy (54.4%). The study also identified some inhibitors to the optimal utilization of the library resources. These<br />

include; lack of awareness and training on the use of the library, low bandwidth, limited access to computer<br />

terminals, etc. The study recommends, among other things, that the bodies responsible for funding the library should<br />

ensure that the Internet bandwidth is upgraded to enhance its speed, users should be trained or given proper<br />

orientation on how to use the resources provided by this library and, the acquisition of more computer systems to<br />

augment the ones provided by MTN foundation<br />

Keywords: digital libraries, electronic resources, library resources use, universities, MTN foundation, Nigeria<br />

1. Introduction<br />

The poor availability and obsolescence of books, journals and other learning materials have always<br />

posed a problem that impairs the ability of higher education institutions in Africa to adequately achieve<br />

their purpose of teaching, research and community service for their expanding population of students<br />

(UNESCO, 2003). Even though Nigeria has more than three hundred higher institutions presently, the<br />

precarious state of the economy has negatively affected the institutions hence scholars and students can<br />

only access a small portion of the necessary library facilities in these institutions. The pitiable state of<br />

library facilities necessitated the Federal Government of Nigeria to make some interventions through<br />

such bodies as, the World Bank Federal Universities Development Sector Adjustment Credit, the<br />

Petroleum Special Trust Fund, National Education Materials Procurement Programme and the Education<br />

Tax Fund. These bodies assisted institutions with the provision of such facilities as computers, books and<br />

journals, and other office equipment for the libraries. Subsequently, at the instance of the former<br />

President, His Excellency, Chief Olusegun Obasanjo, UNESCO facilitated the development of a plan that<br />

sets out a road map to efficiently deliver local and international content to all Nigerian Higher Education<br />

Institutions, staff and students, in the context of their peculiar environments, by means of a Virtual Library<br />

Database provision to augment the existing library resources (UNESCO, 2003). The Virtual Library<br />

Database is meant to be a full text databases covering all fields of study including indigenous content.<br />

Though these steps had some level of positive effect on libraries in Nigeria, the digital divide gap with its<br />

associated problems were still very much evident in Nigerian libraries. It was against this background that<br />

Ogunsola and Aboyade, (2005) observed the urgent need for Nigerian higher institutions to move from<br />

the traditional role of teaching, learning and research to those driven by the information technology. They<br />

concluded that any college or university in Nigeria that does not have full Internet connectivity with<br />

reasonable speed and relevant ICT facilities in the next few years will not be in a position to fulfil the<br />

purposes for its establishment. Ordinarily, given the limited funding available to academic libraries in<br />

Nigeria with the escalating cost of information materials alongside Information Technology (IT) facilities, it<br />

is obvious that the acquisition of these facilities will end up being a day dream if abandoned to be<br />

handled solely by the library. The need to ensure that users of academic libraries in Nigeria have access<br />

to relevant, timely, up-to-date and appropriate information in their fields of interest necessitated the<br />

continued struggle to have in place a digital library that will give users access to electronic information<br />

resources from varying sources.<br />

428


Ngozi Blessing Ukachi<br />

It is in light of the above that MTN Nigeria had in 2005 commenced its Universities Connect project which<br />

provides computers, Internet access and other infrastructures to Nigerian Universities. This project was<br />

embarked upon to achieve MTN’s laudable objectives of bridging the knowledge and digital divide,<br />

enhancing educational infrastructure development and, providing educational resources for effective<br />

learning. However, even though this laudable project has succeeded in making available and also<br />

providing access to diversified digital resources from enormous numbers of libraries globally, there is still<br />

need to evaluate the use and service provision of these libraries vis-a-vis the objectives of establishing<br />

them. This it is believed will expose lapses and give room for amendments where necessary. According<br />

to Nadiri and Mayboudi (2010) library services evaluation help librarians understand users perceptions<br />

and it contributes in improving library service quality and meeting users’ needs better.<br />

2. Objectives of the study<br />

The general objective of this study is to evaluate the use and service provision of these libraries with<br />

regards to the objectives of establishing them.<br />

The specific objectives are:<br />

1. To find out the extent of electronic resources made accessible to users in these libraries<br />

2. To establish if users are given training on how to utilize the resources<br />

3. To determine the extent to which the resources provided by these libraries are being utilized by the<br />

users<br />

4. To find out how the use of the resources of this library has positively impacted on the students<br />

5. To identify factors inhibiting the optimal utilization of the resources provided by these libraries<br />

6. To make recommendations based on the outcome of the study<br />

3. Background information of MTN Nigeria<br />

MTN Nigeria is one of the global mobile telecommunication service providing companies in Nigeria. The<br />

MTN Foundation is the vehicle through which MTN Nigeria implements its corporate social responsibility<br />

programme. The Foundation was established in July 2004 with a focus on giving back to society and<br />

impacting the quality of lives in Nigeria in a meaningful way. The MTN Foundation, which is the<br />

arrowhead of MTN's social investment drive, is playing a key role in deploying various programmes<br />

designed to help reduce poverty and foster sustainable development in Nigeria.<br />

This digital library project in Nigerian Universities by MTN is popularly referred to as MTN Foundation<br />

Universities Connect Project. The institution of this project was brought about by the intense need to<br />

bridge the digital and knowledge divide problem being experienced in the nation by majority of the<br />

citizens who even though they attend Universities, cannot boast of having utilized electronic information<br />

resources in their research activities. This innovative project is designed to provide digital access to<br />

information and infrastructure to Federal Universities in Nigeria. The project facilitates access to a<br />

collection of digital resources from over 5,500 libraries to assist university students and lecturers with<br />

their research works. The project is implemented in partnership with Netlibrary Nigeria Limited. The<br />

present beneficiary institutions of the Universities Connect project are:<br />

For Phase 1- University of Lagos, Akoka, Lagos State<br />

For Phase 2- Ahmadu Bello University, Zaria, Kaduna State<br />

For Phase 3- University of Nigeria Nsukka, Enugu State<br />

For phase 4- University of Benin, Benin City, Edo State. (This particular one has been approved but<br />

not yet commissioned)<br />

MTN Nigeria’s specific objectives of establishing this Education Portfolio are as follows:<br />

To empower through the provision of information and technology resources<br />

To enable skills acquisition, transfer and development<br />

To raise national literacy levels<br />

To provide educational resources for effective learning<br />

To enhance educational infrastructure development<br />

429


The scope of the project is as follows:<br />

Ngozi Blessing Ukachi<br />

128 networked computers, 3 servers, 2 high capacity printers and 100KVA generator<br />

VSAT equipment & internet connectivity bandwidth with 2 years subscription<br />

2 years subscription to electronic resources through the NetLibrary network (Journals, Books, etc)<br />

Conducive study environment through space renovation, provision of adequate lighting, furniture, and<br />

alternative power supply 75b<br />

Technical training for 12 members of existing library staff to work with NetLibrary over 2 years to build<br />

capacity<br />

One-week library awareness to be held annually for students and lecturers<br />

One-week awareness workshops to be held annually for approximately 600 students and 120<br />

lecturers<br />

Initiate a 5–year maintenance contract with the University to ensure a conducive environment with a<br />

clear sustainability strategy<br />

Two years comprehensive insurance cover to take care of theft & fire<br />

An interactive students website - “www.universitiesconnect.com”<br />

4. Methodology<br />

A descriptive research approach was adopted for this study. The purpose of descriptive research is to<br />

describe the current state of affairs at the time of the study without manipulating the variables or having<br />

control over them (Salkind, 2006). The study employed survey procedure for the collection of data.<br />

According to Nworgu (1991) survey method of investigation is one in which a group of people or items<br />

are studied by collecting and analyzing data from only a few people or items considered to be<br />

representative of the entire group. The choice of this research design is considered adequate for the<br />

study because it will enhance and give a better understanding of the situation on ground in relation to the<br />

extent of services provided by the library and also, extent of usage by the users.<br />

This study which was initially programmed to evaluate two out of the three presently existing functional<br />

digital libraries established by MTN Nigeria, ended up evaluating only one, situated at University of<br />

Lagos. This is because the second library at the time of this study was temporarily locked up as a result<br />

of some logistic issues that are yet to be resolved between the institution and the Netlibrary. The<br />

population of this study comprises both the librarian and users of this library.<br />

A self-developed questionnaire tagged Electronic Resources Availability, Extent of Use, Impacts and<br />

Inhibitors (ERAEUII) scale for users was used for data collection. The questionnaire is divided into four<br />

parts. Part A deals with the extent of electronic resources available in the library. Part B concentrated on<br />

the extent of use of this library and its resources. Part C sought responses on the positive impacts of<br />

utilizing this library while Part D concentrated on identifying inhibitor to the optimal utilization of the<br />

resources provided in this library. To complement the questionnaire, the researchers also interviewed the<br />

librarian in-charge of this library. The data collected was subjected to descriptive statistical analysis using<br />

percentages and frequency counts.<br />

5. Findings of study<br />

Questionnaire distribution pattern<br />

A total of One thousand (1000) questionnaires were distributed to the library users. Out of this number,<br />

852 were returned while 812 questionnaires were found usable.<br />

Table 1: List of electronic resources available in this library<br />

1 E- RESOURCES/ DATABASES<br />

2 Sabinet Online Reference Database<br />

3 EBSCOhost Reference Database<br />

4 Jstor<br />

5 HINARI<br />

430


Ngozi Blessing Ukachi<br />

6 MIT<br />

7 SSRN<br />

8 Nation <strong>Academic</strong> Press Database<br />

9 E-Print in physics, Mathematics, computer science and Quantitative Biology<br />

10 Digital Library for earth system education Online<br />

11 Virtual Libraries online<br />

12 Science Direct Online<br />

13 Law Journals Online<br />

14 Oxford Journals Online<br />

15 Ebooks Online Library<br />

16 Linguistics Database Online<br />

17 Guteng Online Library<br />

18 The Encyclopedia<br />

19 The Wiki Service<br />

20 Free Library Online<br />

21 Google Search Engine<br />

22 E-Books Collection<br />

23 File Management<br />

6. Training on the use of the library<br />

On the issue of whether the users are given training on how to use the resources, they all unanimously<br />

answered no. This implies that they are not given any form of training on the use of this library. As a<br />

result of this, majority of the users of this library are only aware of the existence of a few of the database,<br />

thereby underutilizing the resources.<br />

During the interview with the librarian in charge of this library, the researcher was made to know that the<br />

library has not actually been training the users on the use of this library neither have they been holding<br />

one-week library awareness as indicated in the scope of the project by MTN foundation. The reason for<br />

this according to him was that the library lack adequate manpower to be able to perform that function as it<br />

has just four (4) regular staff members.<br />

7. Extent of use of the library<br />

To determine the extent to which the resources provided by these libraries are being patronized by the<br />

users, they were asked to indicate how frequently they utilize the resources in this library. Options such<br />

as; “daily, 2-3 times weekly, weekly, and once in a while”, were given. Their response is represented in<br />

the pie chart below:<br />

Table 2 indicates that 66.7% of the respondents responded that they utilize the library once in a while, the<br />

other 33.3% responded that they utilize it 2-3 times weekly, while none of the respondents attested to be<br />

utilizing it daily or weekly.<br />

Table 2: Response on extent of use of the library<br />

Extent of use Frequency count Percentage<br />

Daily 0 0.0<br />

2-3 times weekly 268 33.3<br />

Weekly 0 0.<br />

Once in a while 544 66.7<br />

431


Ngozi Blessing Ukachi<br />

The librarian was equally asked during the interview to comment on his observation regarding the<br />

utilization of the resources of this library. It was gathered from his response that the library is not being<br />

adequately utilized.<br />

Table 3 illustrates the ways and extent to which the use of the resources provided by this library has<br />

benefitted the students. The findings revealed that 622 (76.6%) of the respondents indicated that the use<br />

of the resources in this library has enriched their awareness on the existence of various kinds of e-<br />

resources. This was closely followed by 559 (68.8%) acceptance of respondent that the use of the<br />

resources has assisted them greatly in doing their assignments. While the option that ‘resources from this<br />

library have made my research work very easy’ received 458 (54.4%) acceptance, ‘my general academic<br />

performance has improved as a result of using MTN Digital library resources in augmenting my<br />

classroom works’ received 320 (39.4%) respondent’s acceptance.<br />

Table 3: Response on positive impacts of utilizing MTN digital libraries<br />

POSITIVE IMPACTS Frequency count Percentage<br />

Enriched my awareness on the existence of various kinds of e- resources 622 76.6<br />

Assisted me greatly in doing my assignments 559 68.8<br />

Resources from this library have made my research work very easy 458 54.4<br />

My general academic performance has improved as a result of using<br />

MTN Digital library resources in augmenting my classroom works<br />

320 39.4<br />

Table 4 shows factors inhibiting the optimal utilization of the resources provided by this library. This<br />

revealed low bandwidth attestation by 786 respondents, which is 96.7% of the study’s user population. A<br />

number of 527 (64.9%) of the respondents indicated that limited access to computer terminals is one of<br />

their inhibitors to optimal utilization of MTN Digital library resources even as Poor attitude of library staff<br />

was indicated by 386 (47.5%) respondents. Irregular power supply received 278 (34.2%) indication by the<br />

respondents while insufficient IT knowledge received 267 (32.8).<br />

Table 4: Response on inhibitors to optimal utilization of MTN digital library<br />

Frequency count Percentage<br />

Inhibitors<br />

Low bandwidth 786 96.7<br />

Irregular power supply 278 34.2<br />

Poor attitude of library staff 386 47.5<br />

Limited access to computer terminals 527 64.9<br />

Insufficient IT knowledge 267 32.8<br />

8. Discussion of findings<br />

The findings of the study show that the necessary information communication and technology facilities in<br />

the form of computers, printers, scanners, Internet access, database, etc, required for service provision in<br />

this library are all available. It also revealed that these facilities are in good working condition. From<br />

personal observation, the library has cconducive study environment. There is adequate space and<br />

furniture provision and also alternative power supply.<br />

On the subject of positive impacts of using the library, high percentage (76.6%) of the respondents<br />

attested that the use of this library have enriched their awareness on the existence of various kinds of e-<br />

resources. A number of 559 respondents being 68.8% of the user population stated that it has assisted<br />

them greatly in doing their assignments while 54.4% confirmed that resources from this library have<br />

made their research work very easy. This result shows that adequate utilization of these resources by<br />

students will enhance their academic performance and assist in making them assets to the society. This<br />

finding corroborates with Tella, Tella, Ayeni and Omoba’s (2007) findings in their work on Self Efficacy<br />

and Use of Electronic Information as Predictors of <strong>Academic</strong> Performance. This work identified that<br />

electronic resources has many functions and benefits that are capable of positively influencing the<br />

academic performance of both students and researchers in the university as well as other higher<br />

educational institutions.<br />

432


Ngozi Blessing Ukachi<br />

On the other hand, the analysis of the results reveals some shortfalls which if not taken care of could<br />

affect the actualization of the goal of establishing this library. The entire user population of the study<br />

unanimously stated that no form of training is given them on the use of this library and its resources. The<br />

repercussion could be seen from the result of the extent of use of the library where not even a single user<br />

accepted to have been using the library on a daily basis rather a very high percentage (66.7%) of the<br />

user population indicated that they use it “once in a while”. It is a common knowledge that an individual<br />

will utilize facilities more when he possesses adequate skills necessary for their use while on the other<br />

hand he cannot effectively utilize what he does not have any training on.<br />

Low bandwidth and limited access to computer terminals were identified as the major inhibitors to optimal<br />

utilization of this Library. This finding agrees with those of Aduwa-Ogiegbaen and Iyanu (2005), Edom<br />

(2007) and Ukachi (2008) who noted inadequate number of ICT facilities, limited access, among other<br />

factors as inhibitors to optimal utilization of this kind of libraries. The forgoing indicates the slow pace at<br />

which the movement towards bridging the digital divide is being made. Poor attitude of library staff was<br />

also identified as another factor inhibiting the optimal utilization of these resources. This may perhaps be<br />

caused by the problem of inadequate staffing being experienced by the library which is likely to be<br />

affecting their service provision.<br />

A survey research on the use of ICT equipment carried out by Daniel et.al (2003) shows 70%<br />

respondents’ acceptance of incessant electric power failure as a major impediment while this particular<br />

work has 34.2% acceptance of this factor. This provision of standby power generator is a pointer that<br />

institutions are now realizing the place of power supply in digital/electronic service provision.<br />

9. Conclusion<br />

This study without doubt has led to the establishment of facts as regards the main objective of the work,<br />

which is to evaluate the use and service provision of MTN Foundation’s University digital library projects<br />

in Nigeria. The result of the study showed that this project is appreciated by users considering their<br />

testimonies concerning the positive impacts of utilizing the resources being provided. At the same time, it<br />

should be noted that mere provision of this resources without creating corresponding awareness on its<br />

existence tantamount to waste of resources as availability is only one of the factors that influence usage<br />

but does not necessarily guarantee optimal usage on its own.<br />

10. Recommendation<br />

Based on the findings of the study, if the digital/electronic libraries must live to continually fulfill their<br />

primary goal which is, the provision of information and services geared at satisfying the information needs<br />

of users, the following recommendations meant to check the usage inhibiting factors should be<br />

considered and implemented.<br />

The bodies responsible for funding the library should ensure that the Internet bandwidth is upgraded<br />

to enhance its speed. This should also be emulated by similar libraries.<br />

To curb the problem of limited access to computer systems, the institution should acquire more<br />

computer systems to augment the ones provided by MTN foundation. Meanwhile, considering the<br />

funding situation of the libraries in Nigeria which Okiy (2005) termed to be inadequate, the library on<br />

its own should also develop some fee-based electronic reference services such as, Selective<br />

Dissemination of Information(SDI), Abstracting and Indexing services to individuals, etc. This will help<br />

them generate funds internally to augment what is being allocated to them for the acquisition of<br />

additional ICT facilities to enable larger number of users to access them.<br />

Users should be trained or given proper orientation on how to use the resources provided by this<br />

library more effectively. The relevance of library orientation should never be underestimated be it in<br />

the traditional or electronic library environment.<br />

To actualize the above, more staff members should be employed in the library to assist in the service<br />

provision of the library.<br />

References<br />

Aduwa- Ogiegbaen, S. E. and Iyanu, E. O. S (2005) Using Information Technology in Secondary Schools in Nigeria:<br />

problem and prospects. Educational Technology and Society. 8(1). P.104-112.<br />

Edom B. O. (2007) An assessment of the problems of the use of ICT facilities in information sourcing and retrieval by<br />

the academic staff in Nigerian universities. The InformationTechnologist: 4(2).p.113-127<br />

433


Ngozi Blessing Ukachi<br />

Daniel, J. O. et al (2003) Forty years of information and communication technology (ICT) Library services to the<br />

nation: in S. O. Lanlokun (ed.) forty years of library services in Nigeria. The Guardian Newspaper, March 24.p.2.<br />

http://www.mtnonline.com/mtn-foundation/...http://www.mtnonline.com/MTN-Foundation/index.asp<br />

Nadiri, H. and Mayboudi, S. M. A. (2010) Diagnosing university students’ zone of tolerance from university library<br />

services. Malaysian Journal of Library & Information Science, 15(1). p.1-21.<br />

http://majlis.fsktm.um.edu.my/document.aspx?FileName=842.pdf. Accessed on 22/01/2011<br />

Nworgu, B. G. (1991) Educational research: basic and methodology. Ibadan, Wisdom Publishing<br />

Ogunsola, L. And Aboyade, W. (2005) “Information and Communication Technology in Nigeria: Revolution or<br />

Evolution” Journal of Social Sciences, 11(1): 7-14<br />

Okiy, Rose B. (2005) Funding Nigerian libraries in the 21 st century: will funding from alternate sources suffice? The<br />

Bottom Line: managing library finances, 18(2). p.71-77<br />

Salkind, N. J. (2006) Exploring research. 6 th ed. Pearson: New Jersey. Samba, A. 2007<br />

Tella, A., Tella, A., Ayeni, C. O. and Omoba, R. O. (2007), Self Efficacy and Use of Electronic Information as<br />

Predictors of <strong>Academic</strong> Performance. Electronic Journal of <strong>Academic</strong> and Special Librarianship. 8(2).<br />

http://southernlibrarianship.icaap.org/content/v08n02/tella_a01.html Accessed on 22/06/2010<br />

Ukachi, N. B. (2008) Utilisation of Information Communication Technologies (ICTs) in Reference Services of<br />

<strong>Academic</strong> Libraries: Threats and Challenges. The InformationTechnologist. 5(2).p62-69<br />

UNESCO, (2003),“Feasibility Study on the Development of a Virtual Library by Institutions of Higher<br />

Education in Nigeria”, Consultancy Support Services Ltd., Abuja, Nigeria, available at:<br />

http://www.portal.unesco.org/en/files/39276/...Final.pdf/FS4VL_Final.pdf. Accessed on 18/11/2009<br />

434


Cloud Computing-Based IT Solutions for Organizations With<br />

Multiregional Branch Offices<br />

Harris Wang<br />

Athabasca University, Canada<br />

harrisw@athabascau.ca<br />

Abstract: One of the most significant phenomena of the new century is globalization. As business goes global,<br />

multiregional branch offices are needed and networked computing and information services must then be established<br />

for those branch offices. In this paper we investigate a cloud computing based approach to the rapid deployment of<br />

computing and information services for organizations with multiregional branch offices. We first take a look at the<br />

general process leading to the deployment of Computer and Information Technology (CIT) services for organizations,<br />

and then present some cloud computing-based solutions for organizations with multiregional branch offices, followed<br />

by discussions about their key features as well as issues and concerns surrounding the proposed IT solutions. An<br />

important contribution of this paper is a generalized view of the cloud computing-based approach, which may be<br />

used as guidance in implementing and deploying such IT solutions.<br />

Keywords: cloud computing, globalization, information technology infrastructure<br />

1. Introduction<br />

In today’s world, the reach of computer and information technology is vast and deep. Computer and<br />

information technology are prevalent in every corner of our society from health care and military<br />

deployments, to that of the corporate business world; using computers and information technology has<br />

become a fundamental part of daily operations for almost all organizations. Information technology drives<br />

business and communications through virtually removing borders and distance limits and making it<br />

possible for instant commerce and communication globally between organizations or internally.<br />

The expectation that surrounds the use of computer and information systems on this global network has<br />

become one that society expects instant communications with instant gratification and results from all<br />

businesses at all times. In order to provide this type of service it is necessary to have a robust and<br />

reliable information technology infrastructure underpinning the communications and information systems.<br />

Traditionally, the information technology infrastructure that supports these systems needed for<br />

businesses with multiregional branch offices usually starts internally with a core network and data center<br />

for the originating site or head office and extends to branch offices worldwide. This traditional approach<br />

often provides more reliable and better guaranteed computing and information services. With the<br />

advances in computing and information technology, we now may have more options in providing the IT<br />

needs to businesses with branch offices worldwide. In this paper, we will explore a cloud computing<br />

based approach to the rapid deployment of computing and information technology services for<br />

businesses with offices worldwide.<br />

2. General management process towards an IT solution for businesses<br />

In general, any new branch office will require a data center, network core and infrastructure,<br />

telecommunications systems, 3-tier application systems, web servers, database servers and client<br />

computers, and all these elements require a high degree of information security considerations to be<br />

taken. No matter how these elements will be implemented, some general Project Management (PM)<br />

steps must be taken (PMI 2004).<br />

In order to determine exactly what IT infrastructure and systems are needed and how to implement the IT<br />

infrastructure and determined systems for a new business expansion, the first step is defining the needs<br />

of the business. In this case a Chief Technical Architect (CTA) or the like will need to work with the<br />

functional representatives from the business and defines what will be needed to support the daily<br />

operations of the business; from an ERP (for Enterprise Resources Planning) system and Email services<br />

to Computer Numerical Controlled (CNC) systems and Computer-Aided Design (CAD) systems, this<br />

should be determined in the project conception stages. This Initiation phase of the project allows the<br />

business to communicate the technical needs or concepts to be evolved.<br />

The next phase of the project is Project Planning (PP), which in the project lifecycle is the next logical<br />

progression for the project. It is important in this phase to determine the full extent of the work packages<br />

or work breakdown structure (WBS), the sequencing of activities, resources required, risks and risk<br />

435


management plans, cost estimating and budget, change and quality control, procurement plans, plan and<br />

hold a formal project kick off meeting and define the finalized project plan (PMI 2004).<br />

The last phase of the project is Project Execution where the work begins to occur is signified by acquiring<br />

the project team, receiving responses from any request for proposal, selection of the vendor based on<br />

proposal, execution of the work packages, perform quality assurance, and the PM distributes and<br />

controls project information and implements administrative closure procedures.<br />

This 3-phase process of project management is applicable when using a cloud computing based IT<br />

solution to meet the computing and information technology needs of businesses with multiregional branch<br />

offices.<br />

3. Cloud computing based solutions<br />

Cloud computing is defined as “a pay-per-use model for enabling available, convenient, on-demand<br />

network access to a shared pool of configurable computing resources (e.g., networks, servers, storage,<br />

applications, services) that can be rapidly provisioned and released with minimal management effort or<br />

service provider interaction” (Pearson 2009). In literature and practice, there are three cloud computing<br />

models: Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service<br />

(SaaS) (Fouquet 2009, Gruman 2008).<br />

In the IaaS model, users subscribe for uses of certain components of provider’s IT infrastructure.<br />

Although the subscribers don’t have control of the entire cloud infrastructure, they do have control over<br />

selected portions of it, such as firewalls, operating system, deployed applications and storage.<br />

In the PaaS model, a combination of applications, which forms a platform, is subscribed as a service by<br />

users. For example, a combination of software tools may be used as a programming and software<br />

platform.<br />

The SaaS model can be seen as special case of PaaS, where a single application can be subscribed as<br />

a service. Such services are often accessed through a Web browser.<br />

As for cloud, there are private cloud, public cloud and hybrid cloud based on the relationship between<br />

cloud owners and cloud users (Buyya 2008, Dikaiakos 2009). Private Cloud is usually owned and used<br />

by the same organization such as a corporation. It often refers to a proprietary computing infrastructure<br />

owned by the organization, and provides computing and information services to its employees behind the<br />

organization’s firewall. Public cloud often refers to computing and IT infrastructure that is owned by an<br />

organization but provide computing and information services to external users or subscribers. By<br />

subscribing services provided by other well established companies, new start-ups, for example, can<br />

quickly realize their computing and information technology needs without investing so much money and<br />

time to implement their own computing and IT infrastructure.<br />

Hybrid cloud is a mixture of public cloud and private cloud. It can be useful for some well established<br />

corporations who already have their own computing and IT infrastructure, but need additional computing<br />

and IT services for expansion in new areas. By subscribing needed computing and information services<br />

available in public cloud, they can make use of the services more quickly without investing big money and<br />

lengthy time on implementing their own CIT infrastructure. Accordingly, we derive from the above the<br />

following cloud computing based solutions for businesses with multiregional branch offices, to meet their<br />

computing and information technology needs.<br />

3.1 Private cloud based solution to meet the IT needs of multiregional branch offices<br />

For organizations who have well established computing and information technology infrastructure, such<br />

as some high-tech companies in the IT business, a completely private cloud based solution may be a<br />

better choice to provide computing and IT services to new branch offices in other cities or other countries.<br />

In such a case, the data centre, servers and all major computing and IT devices reside behind the firewall<br />

on the organization’s enterprise network, located on site of the head office, while users in branch offices<br />

access the computing and information services through VPN or a Web browser, if a Web interface has<br />

been made available for accessing the service. Figure 1 depicts the architecture of our private cloud<br />

based solution.<br />

436


Branch Office A<br />

Branch Office C<br />

Figure 1: Architecture of private cloud based solution<br />

In private cloud based solution since everything is under control of the very same organization, this<br />

solution gives the organization the total freedom and autonomy in managing all the components of the<br />

computing and information technology infrastructure.<br />

3.2 Federated cloud based solution to meet the IT needs of multiregional branch offices<br />

Federated cloud can be seen as a variant of private cloud. Same as in private cloud, in federated cloud<br />

computing and information technology infrastructure is still privately owned by the sole organization, but<br />

the equipment, servers and services are distributed among head office and branch offices. This may be<br />

necessary when different branch offices have different missions and each needs more dedicated<br />

computing and information services. For example, one branch may be working on data mining, while<br />

another branch may be more on server development. The architecture of federated cloud based solution<br />

is showed in figure 2.<br />

Branch Office A Branch Office B<br />

Branch Office C<br />

Figure 2: Architecture of federated cloud based solution<br />

3.3 Public cloud based solution to meet the IT needs of multiregional branch offices<br />

A public cloud based solution is for the organization to subscribe all needed computing and information<br />

technology services from providers in the public cloud. This solution is suitable for organizations that<br />

have no resources or interest to implement their own CIT infrastructure, or small start-ups. By subscribing<br />

the needed computing and IT services readily available in the public cloud, a start-up can quick get its<br />

business going and have the creative ideas tested. If the business doesn’t fly, it can easily get off the<br />

boat with less to lose. Figure 3 depicts the architecture of public cloud based solution.<br />

Figure 3: Architecture of public cloud based solution<br />

Private cloud owned by and hosted<br />

at the head office of the cooperation<br />

Branch Office E Branch Office F<br />

Branch Office G<br />

Branch Office A<br />

Branch Office C<br />

Branch Office E<br />

Public Cloud<br />

Head Office of the<br />

Cooperation<br />

437<br />

Branch Office D<br />

Branch Office B<br />

Branch Office D<br />

Branch Office A<br />

Branch Office D<br />

Branch Office F


3.4 Hybrid cloud based solution to meet the IT needs of multiregional branch offices<br />

As in the common definition, hybrid cloud involves both private and public cloud. A hybrid cloud based<br />

solution may be applicable for an organization that has an established CIT infrastructure sufficient for its<br />

current needs, but doesn’t want to invest big money and time to expand its current CIT infrastructure for<br />

new business. In this case, it would rather choose to get the needed CIT services from a reliable source<br />

in the public cloud. By doing this, it may be more easily turn around if the new adventure doesn’t go well.<br />

The architecture of hybrid cloud based solution is shown in figure 4.<br />

Branch Office A<br />

Private<br />

Branch Office A<br />

cloud owned<br />

by and<br />

hosted at<br />

Branch Office A<br />

the head<br />

office of the<br />

cooperation Branch Office A<br />

Branch Office A<br />

Figure 4: Architecture of hybrid cloud based solution<br />

Public Cloud<br />

4. Advantages and challenges of cloud computing based solution<br />

Cloud computing based IT solution for organizations with multiregional offices brought both advantages<br />

and challenges. The advantages are as follows:<br />

Cloud computing provides organizations with more agile solutions to meet their IT needs. With cloud<br />

computing, they can quickly get the IT services needed to run their business. This is especially<br />

suitable for new business with a brilliant but risky adventure. Because the new business idea is<br />

brilliant, they need to get the idea tested as soon as possible before it is stolen by competitors;<br />

because it is a risky adventure, it is very likely to fail. By subscribing CIT services available from the<br />

public cloud, they could avoid big loss in CIT infrastructure investment (Cohen 2004).<br />

By subscribing services from a third-party, an organization can save money on capital investment. It<br />

can also save on maintenance because the CIT infrastructure is owned and maintained by the thirdparty.<br />

Compared to CIT infrastructure developed from scratch, cloud computing based solutions can be<br />

more reliable by subscribing well established computing and information technology services<br />

provided by trusted third-parties (Stantchev 2009).<br />

Because additional CIT services can be easily subscribed from providers in the cloud, a cloud<br />

computing based IT solution can be more scalable. It can also be more easily scale down if<br />

necessary, because unsubscribing unneeded services are usually less painful than disposing some<br />

expensive CIT equipments (Buyya 2009, Brandic 2009).<br />

Generally speaking for the entire IT industry, cloud computing provide better IT solutions to utilize IT<br />

infrastructure and equipments, as these often expensive infrastructure and equipments can now be<br />

shared among more users. This also means the overall power consumption will be lower, and hence<br />

the carbon footprint will be smaller. So, it is good for the environment too (Buyya 2008).<br />

In cloud computing based solutions, because important data centers and servers are often owned<br />

and managed by big well established companies who usually have much better expertises and<br />

resources to secure their equipments and services, cloud computing based solutions are often<br />

secure in this regard (Stantchev 2009).<br />

There are few challenges or concerns surrounding could computing based IT solutions. There challenges<br />

or concerns are:<br />

1. The operational costs may be high. We mentioned that the capital cost can be much lower when a<br />

cloud computing based solution is chosen. However, its operational cost may be higher than running<br />

its own CIT infrastructure because it may be required to pay big money for the subscribed services.<br />

438


This all depends on the relationships and contract between the organization and the services<br />

provider or providers (Li 2009).<br />

2. Reliability may be questionable if the CIT services provided in the cloud are not carefully chosen.<br />

In theory, a cloud computing based IT solution should provide reliable CIT services if the providers<br />

are carefully chosen. There are two factors that may cause the CIT services obtained from cloud<br />

computing unreliable. These factors include:<br />

a. Unpredictable failure of network external to the organization. Since those portions of network are<br />

not under control of the organization, it cannot really do anything to avoid such failure, or to ensure a<br />

quick recovery once such a failure has occurred.<br />

b. Services broken down at the service provider. It is very important to choose a reliable and trusted<br />

service provider. Even if this has been done, there is still no guarantee as things may change over<br />

time. The provider may have gone bankrupt; it might be sold to another party which is unfamiliar or<br />

even hostile to the subscriber. The original owner of the service provider may turn to be unfriendly for<br />

some reasons.<br />

3 .In a cloud computing based solution, the security of data and subscribed CIT services are in thirdparties’<br />

hand. They care secure only if the third parties are doing their good job. However, even if<br />

great effort has been taken in choosing the service providers, security may still be a problem fir the<br />

following reasons (Pearson 2009, Kaufman 2009, Kandukuri 2009):<br />

a. They may be an oversight when evaluating the security measures implemented at the service<br />

providers;<br />

b. There may be new threats that the service providers are unprepared for. If the providers do not<br />

implement anti-threat measures in time and effectively, the security of the data and CIT services may<br />

be at great risk;<br />

c. The service providers may change their security practice or even side (due to pressure from the<br />

top, for example).<br />

d. Because in a cloud computing based solution, users access data and subscribed services through<br />

the Internet, security breach may occur during transmission between users and the service providers.<br />

5. Conclusions<br />

Cloud computing brings us both opportunities and challenges. With the advancement of computer and<br />

network technologies, more reliable and powerful computing and information technology services have<br />

become available on the Internet, and high speed reliable internet access to these services is becoming a<br />

reality for many users in the world. Therefore, it is possible for organizations with multiregional branch<br />

offices to get the needed computing and information technology services through cloud computing.<br />

In this paper, we presented four cloud computing based solutions: private cloud based, federated cloud<br />

based, public cloud based, and hybrid cloud based. The subtle differences between these solutions, as<br />

shown in the figures, are important because they have crucial impact to the details of implementation and<br />

even administration, including accounting and security management. For example, in the private or<br />

federated cloud computing based solutions, the accounting can be simple because there is no external<br />

body involved, but the system administration may be heavy especially in the case of federated cloud<br />

computing because of the equipments and services are distributed amongst the branch offices. This may<br />

increase the complexity of the entire system. In case of public cloud computing based solution, the<br />

organization may have less to do in security, but the accounting will have to consider the operational cost<br />

of subscribed CIT services.<br />

We also pointed out in this paper some advantages of using cloud computing based solutions, as well as<br />

some issues. It is important for an organization to carefully study both these advantages and issues when<br />

considering a cloud computing based solution for its needs.<br />

References<br />

Buyya, R., Yeo, C. S. and Venugopal S. (2008) Market-Oriented Cloud Computing: Vision, Hype, and Reality for<br />

Delivering IT Services as Computing Utilities. Department of Computer Science and Software Engineering, The<br />

University of Melbourne, Australia. pp. 9.<br />

Buyya, R., Ranjan, R. and Calheiros, R.N. (2009) Modeling and simulation of scalable Cloud computing<br />

environments and the CloudSim toolkit: Challenges and opportunities, International <strong>Conference</strong> on High<br />

Performance Computing & Simulation, HPCS '09. Page(s):1 – 11<br />

439


Harris Wang<br />

Brandic, I. (2009) Towards Self-Manageable Cloud Services, Annual IEEE International Computer Software and<br />

Applications <strong>Conference</strong>, 2009. COMPSAC '09. 33rd Volume 2, Page(s):128 - 133<br />

Cohen, D., Lindvall, M. and Costa, P. (2004) An introduction to agile methods, Advances in Computers, New York,<br />

Elsevier Science, pp. 1-66.<br />

Dikaiakos, M.D., Katsaros, D., Mehra, P. Pallis, G. and Vakali, A. (2009) Cloud Computing: Distributed Internet<br />

Computing for IT and Scientific Research, Internet Computing, Issue 5, Volume 13, IEEE.<br />

Gruman, G. (2008) What cloud computing really means. InfoWorld (http://www.infoworld.com), retrieved 2010-01-17<br />

from http://www.infoworld.com/d/cloud-computing/what-cloud-computing-really-means-031<br />

Fouquet, M. and Niedermayer H., Carle G. (2009) Cloud computing for the masses, in Proceedings of the 1st ACM<br />

workshop on User-provided networking: challenges and opportunities, pages 31-36.<br />

Kandukuri, B.R., Paturi, V.R. and Rakshit, A. (2009) Cloud Security Issues, IEEE International <strong>Conference</strong> on<br />

Services Computing, SCC '09. Page(s):517 – 520<br />

Kaufman, L.M. (2009) Data Security in the World of Cloud Computing, Security & Privacy, IEEE, Volume 7, Issue<br />

4, Page(s):61 – 64<br />

Li, X., Li Y., Liu T., Qiu J. and Wang F. (2009) The Method and Tool of Cost Analysis for Cloud Computing, IEEE<br />

International <strong>Conference</strong> on Cloud Computing, CLOUD '09. Page(s):93 – 100<br />

Pearson, S. (2009) Taking account of privacy when designing cloud computing services, Proceedings of the 2009<br />

ICSE Workshop on Software Engineering Challenges of Cloud Computing, p.44-52.<br />

Project Management Institute (2004) Guide to the Project Management Body of Knowledge, 3rd Edition.<br />

Stantchev, V. (2009) Performance Evaluation of Cloud Computing Offerings, Third International <strong>Conference</strong> on<br />

Advanced Engineering Computing and Applications in Sciences, ADVCOMP '09, Page(s):187 – 192<br />

440


Case Study on Information Evaluation by GIS for Aging<br />

Society Urban Planning: Information Evaluation of Shrinking<br />

Cities<br />

Yupeng Wang, Hiroatsu Fukuda and Kiyoshi Shinriki<br />

The University of Kitakyushu, Japan<br />

y09e0101@hibikino.ne.jp<br />

fukuda@env.kitakyu-u.ac.jp<br />

shinriki@hq.kiu.ac.jp<br />

Abstract: An understanding of the sites upon which they are working is crucial to city planners at every stage of the<br />

urban design process. Usually, the information (such as population, landform, building conditions, and economic<br />

data) is too complex to be processed and reported for analysis through the ordinary statistical approaches. The GIS<br />

(Geographic Information System) offers software that can collect related databases for different fields onto one map,<br />

allowing both the individual analysis diagram and the compound analysis to be easily achieved. Thus, GIS provides<br />

planners a more comprehensive understanding of the site upon which they are working. Our research is an actual<br />

urban design project in Japan, an example of the utilization of GIS in urban planning. It is a typical case, with an<br />

intricate landform and typical population composition. First, GIS analyzes the basic data, covering landform,<br />

construction age, and population. These analyses occur in two stages: first, a comparative study of the population<br />

structure over the last 10 years in different landform areas, then the forecasting of its evolution. Using these<br />

analyses, the proposal plan tries to improve the residential environment and prevent decline by retaining the local<br />

population and attracting outsiders. The introduction of GIS allows digital data to be composited with draft data,<br />

resulting in clearer and more precise analyses. This method produced accurate dates for Yahatahigasi-ku, therefore<br />

assisting the evaluation of data reflecting the current situation and orienting the next step of the urban design. It has<br />

thus been demonstrated that GIS enhances the efficiency of site analysis. We consider this study of GIS in urban<br />

planning to be a development of GIS application. This paper presents a new information evaluation method for urban<br />

planning that provides more social information on the target area than previous methods and shares the results of<br />

information evaluations.<br />

Keywords: urban design, GIS, population, landform, building state<br />

1. Introduction<br />

New urban proposal plans usually become necessary as a city develops, in order to deal with variations<br />

in population and transformations in lifestyles. Studying the site from different perspectives and<br />

developing a comprehensive understanding of the site are the key steps to a successful plan. The GIS<br />

can help planners gather both the abstract and concrete data into a single graphic. This paper proposes<br />

the introduction of the GIS in the site analysis stage. It discusses the GIS analysis method through a case<br />

study of Yahatahigasi-ku, Japan. Based on these analyses, the proposal plan tries to improve the<br />

residential environment and prevent urban decline by retaining the local population and attracting<br />

outsiders.<br />

2. Goal, research objective, and methods of research<br />

2.1 Goal<br />

Japan’s aging society is becoming a serious problem, which must be analyzed from different<br />

perspectives if it is to be understood and solved. This paper aims to find a new approach that involves the<br />

GIS in the analysis stage.<br />

2.2 Research objective<br />

This study is an urban regeneration proposal plan for Yahatahigasi-ku, Japan. This area became the<br />

biggest industrial area of Japan after the Second World War but is now struggling against depopulation<br />

and an aging community. Furthermore, the houses built during Japan’s industrial age are becoming<br />

decrepit, and most of the land lies on a steep slope. The proposal plan explores ways of solving these<br />

problems and providing residents with a better living environment.<br />

Landform<br />

441


Yupeng Wang et al.<br />

To accommodate employees, dwellings were built on the mountain around the industrial area. As<br />

employees aged, the inconvenient landform made it difficult for them to access the facilities near their<br />

houses. Young people are moving away in search of more comfortable residential environments, while<br />

few people move in (though the local population is increasing). Meanwhile, industrial styles have<br />

changed: factories are smaller, and some of the flat areas can be redeveloped.<br />

Construction age<br />

Dilapidated dwellings in this area also pose a problem. Traditional wood dwellings constitute the majority<br />

of the street. The number of empty houses is increasing, though, especially in the steeply sloped area.<br />

An accurate average date of dwelling construction would enable an analysis of the ward’s actual<br />

condition.<br />

Population<br />

Yahatahigasi-ku had its glory days during the development of the West Japan Steel Factory from 1950 to<br />

1960. The city’s population was at its highest at that time. However, as the industry innovated, local<br />

residents began to leave. The current population of 60 thousand will decrease to 20 thousand in about 40<br />

years. Though depopulation and collective aging are problems throughout Japan, the problem in<br />

Yahatahigasi-ku is particularly serious.<br />

2.3 Methods of research<br />

This paper uses GIS to collect data on landform, construction dates, and population in an analysis of the<br />

ward’s actual condition and future. This is a new approach to site analysis.<br />

3. Data analysis through GIS<br />

3.1 Landform<br />

Population affects the size of city. As a city grows, dwellings and other buildings are built. Given a<br />

haphazard layout, a residential environment becomes inconvenient, energy is wasted, and social<br />

problems emerge. A contour line with a 2m pitch and building layout are put into one picture (see<br />

Figure1), showing the connection between the landform and buildings. Factorial facilities are built in the<br />

beach area. A great number of dwellings are on the steeply sloped area of the mountain. This condition<br />

flows from the industrial age, when the city grew through a rapid increase in population. It is necessary to<br />

remediate access methods for the elderly in a sustainable urban design.<br />

Figure 1: Landform and building distribution in Yahatahigasi-ku (2005)<br />

442


Yupeng Wang et al.<br />

To express the steepness of these areas, “slope degree” is introduced as a standard of measure.<br />

Differences in height and average slope in each 100m×100m mesh can be calculated by counting the<br />

number of contour lines in each mesh area. As shown in Figure 2, the slope degree is the same as the<br />

number of contour lines.<br />

Figure 2: Definition of “slope degree” in each mesh<br />

Figure 3: Image of layer overlapping<br />

Using the overlapping mesh layer and contour line layer seen in Figure 3, we found the slope degree of<br />

each mesh. Consequently, the overlap of mesh upon building layer and the slope degree of each building<br />

can be expressed in different colors according to the different slope degrees (as shown in Figure 4).<br />

Slope degree of each mesh → Layer overlapping → Slope degree of each building<br />

Figure 4: Explanation of building slope degree<br />

443


Yupeng Wang et al.<br />

Figure 5 shows that some buildings have a low slope degree in the high-level area, namely those<br />

buildings built on the flat area of the mountain. A comparison with Figure1 shows that buildings closer to<br />

the top of mountain have a higher slope degree. Poor approach planning makes daily life inconvenient for<br />

city dwellers.<br />

Figure 5: Slope degree of each building<br />

3.2 Construction age<br />

Unlike the stone masonry buildings of Europe, the traditional wooden dwellings of Japan last about 30<br />

years, after which they become decrepit. Moreover, during the immediate postwar period of 60 years ago,<br />

a lack of resources in Japan reduced the quality of Japanese building materials and shortened the life<br />

spans of Japanese dwellings.<br />

In Yahatahigasi-ku, the percentage of dwellings more than 30 years old is higher than 60% in almost all<br />

of the towns. In most towns, 80% of the dwellings are more than 30 years old. In about half the towns,<br />

20% of the dwellings are older than 60. The layout showing construction age appears in Figure 6. A great<br />

number of old buildings lie in the steep slope area, some of which have been abandoned by their owners<br />

and remain empty. The number of empty dwellings will increase as the population decreases.<br />

The average construction age of towns is shown in Figure 7. Almost all of the towns are over 30 years<br />

old. Significantly, the average construction age of some towns is over 60, implying that the whole street is<br />

dilapidated and that a macro-scale reconstruction conforming to the needs of modern lifestyles is needed<br />

here.<br />

3.3 Population<br />

Local activity creates a bustling community and conditions the atmosphere of the street. Thus, population<br />

and the construction of population affect local revitalization. Population layout and construction should be<br />

analyzed before urban planning begins.<br />

As shown in Figure 8, the population density of towns on the sloped area is lower than that in the flat area<br />

because most of dwellings on the slope are isolated houses. Apartments have recently been built close to<br />

the railway station. The population of the flat area is increasing slowly, but the total population of this<br />

ward is still decreasing, especially in the towns on the steeply sloped area.<br />

444


Yupeng Wang et al.<br />

Figure 6: Construction age of buildings in Yahatahigasi-ku<br />

Figure 7: Average construction age of each town<br />

An aging society is that in which the percentage of the population over 65 is higher than 25%. A<br />

percentage higher than 50% produces a limitative colony, a social community that is difficult to maintain.<br />

As Figure 9 shows, almost all of the towns satisfy the criterion of an aging society. The percentage of the<br />

population that older than 65 is higher on the steeply sloped area, where it is even higher than 50% (thus<br />

meeting the criterion of a limitative colony). All of the limitative colonies are situated in the highly sloping<br />

area of the mountain.<br />

445


Yupeng Wang et al.<br />

Furthermore, as Figure10 shows, the percentage of the population older than 75 is higher than 20% in<br />

many towns, suggesting that collective aging is a major problem in this ward.<br />

Figure 8: Population density of Yahatahigasi-ku (2008)<br />

Figure 9: Percentage of population older than 65 (2008)<br />

446


Figure 10: Ratio of population older than 75 (2008)<br />

Yupeng Wang et al.<br />

We discover future population and population construction trends and predict the ward’s population size<br />

in 10 years (2018) by applying an equation to the trend line of the last 10 years (1998 ~ 2008).<br />

P18=P18T-(P08N-P08)×P18T/P08T<br />

Where<br />

P18:The population of 2018 (prediction)<br />

P08:The population of 2008<br />

P18T:Transitive the population of 2008 to 10 years later.<br />

P08T:Transitive the population of 1998 to 10 years later.<br />

Figure 11: Population of each age in Yahatahigasi-ku (1998)<br />

447


Yupeng Wang et al.<br />

Figure 12: Population of each age in Yahatahigasi-ku (2008)<br />

Figure 13 shows the population size 10 years from now, based on the trend of the last 10 years. By then,<br />

the population will have decreased from 73162 to 63045. The percentage of the population that older<br />

than 65 is going to increase from 30.2% to 35.4%, far higher than the criterion of 25% that defines an<br />

aging society.<br />

Thus, if current conditions continue, the population is going to decrease, and the percentage of elderly is<br />

going to increase. Some countermeasures should be introduced to this area immediately to maintain its<br />

vitality.<br />

Figure 13: Population of each age in Yahatahigasi-ku (2018-prediction)<br />

(The date in the dark color means the population is higher than a thousand.)<br />

4. Conclusion<br />

This paper discusses an urban regeneration proposal plan for Yahatahigasi-ku as a case study and<br />

proposes the GIS method of site analysis. The GIS information evaluation of Yahatahigasi-ku reveals that<br />

many dwellings in the highly sloped area suffer from a poor living environment because of their advanced<br />

ages. Poor approach planning forces inhabitants to walk everywhere, which is hard for the elderly who<br />

constitute a high proportion of Yahatahigasi-ku’s population.<br />

The introduction of GIS allows digital data to be composited with draft data, resulting in clearer and more<br />

precise analyses. This method produced accurate data for Yahatahigasi-ku, therefore assisting the<br />

evaluation of data reflecting the current situation and orienting the next step of the urban design. It has<br />

thus been demonstrated that GIS enhances the efficiency of site analysis.<br />

Acknowledgments<br />

The authors wish to express their appreciation to “Revitalization of Local Area Project” for providing the<br />

subvention for this research.<br />

References<br />

Dewanchker B., (1997) “Comparison of the Change in Industrial Development in the Kitakyushu Industrial Zone and<br />

the EMSCHER Zone in Germany”, Journal of architecture, planning and environmental engineering.<br />

Transactions of AIJ (502) pp.51-56<br />

Dewanchker B., Takahasi N. and Ojima T., (1997) “Ecological Improvement and the Redevelopment of an Abandon<br />

Industrial Site in the City of Uozu”, Summaries of technical papers of Annual Meeting Architectural Institute of<br />

Japan. pp.821-822<br />

Fukahori H. (2006), “Disaster prevention map making with IT method”, Journal of Japan Sewage Works Association,<br />

43(527) pp.43-46<br />

Fukahori H. (2007), “Make Kitakyusyu-city an nature environment capital”, Journal of sewerage, monthly 30(2) (422)<br />

pp.38〜41<br />

Gao W., Ariyama T. and Ojima T., (1999) “Prediction on the Energy Consumption of Renewing the Residential<br />

Recycle Material”, Journal of architecture, planning and environmental engineering. Transactions of AIJ (516)<br />

pp.101-106<br />

448


Yupeng Wang et al.<br />

Gao W. and Ojima T., (1995) “Study on Planning an Ecological City in SHITAMACHL Tokyo”, Journal of architecture,<br />

planning and environmental engineering. Transactions of AIJ (472) pp.63-71<br />

Gao W., (1993) “Thermal Effects of Open Space with a Green Area on Urban Environment: Part I: A theoretical<br />

analysis and its application”, Journal of architecture, planning and environmental engineering. Transactions of<br />

AIJ (448) pp.15-27<br />

Song H., Guan D., Fukahori H. and Gao W., (2009) “Study on utilization of semi-industrial districts based on GIS in<br />

Kitakyushu : Investigation of present status of semi-industrial districts in Kitakyushu Part 1”, Journal of Asian<br />

Architecture and Building Engineering, pp.529-532<br />

Urban planning general overview of Kitakyusyu city of Japan, (2007) Planning department in city urban planning<br />

bureau of Kitakyusyu, Japan<br />

449


The Impact of Software Test Constraints on Software Test<br />

Effectiveness<br />

Grafton Whyte and Donovan Lindsay Mulder<br />

University of the Western Cape, Cape Town, South Africa<br />

drgwhyte@aol.com<br />

donovanmulder@hotmail.com<br />

Abstract: Software testing is the one of the primary methods used in the validation and verification of output in the<br />

software development industry. It is seen as a key method for achieving software quality, reliability, fitness for<br />

purpose and customer satisfaction. Software testing is however an expensive process accounting for as much as<br />

50% of the cost of developing software based systems. In recent years, software testing as a discipline has come<br />

under pressure due to time, cost and skills constraints. These constraints impact negatively upon software test<br />

effectiveness. Therefore it is critical to identify and implement test tools that reduce the negative impact of software<br />

test constraints on software test effectiveness. In this paper the researcher examines some of the most popular<br />

software testing tools such as test case prioritisation, test suite reduction and test selection criteria, to identify: Which<br />

individual test tools are most likely to yield optimal test effectiveness and, Which combination of test tools is most<br />

likely to yield optimal test effectiveness and mitigate the effect of test constraints An extensive review of the software<br />

testing literature was conducted and used to construct a survey instrument as the basis for examining the impact of<br />

test constraints on software test methodology. The survey was issued to expert software test practitioners from<br />

various locations globally; the sample consisted of 43 test cases. The main findings were that no one approach to<br />

testing would yield satisfactory results but a combination of two or more test types from Automated testing, Smoke<br />

testing, Test case prioritisation and Regression test selection could yield effective software testing results and<br />

mitigate the effects of test constraints.<br />

Keywords: software test tools, software test effectiveness, software test constraints, test selection methodology, test<br />

case selection criteria<br />

1. Introduction<br />

As the significance of software increases aspects such as quality, reliability and customer satisfaction<br />

have become strategic goals for software development organisations (Chin-Yu, 2005). Software testing is<br />

a costly (Bryce and Colbourn, 2005) and unavoidable task (Bertolino, 2007). It is complex and arduous<br />

and can consume more than 50% of the cost of a software development project (Shahamiri, et al., 2009;<br />

McMinn, 2009; Srikanth and Williams, 2005). Engel and Last (2007) state that inadequate execution of<br />

software and systems verification, validation and testing account for losses that can eclipse more than<br />

10% of a company’s turnover.<br />

Software testing has been the method of choice for software validation and verification (Bertolino, 2007)<br />

and remains the method of choice through which confidence in the end product is realised (Ramler and<br />

Wolfmaier, 2006). In today’s software development environment testing has come under pressure due to<br />

shorter time to market, shrinking budgets and higher quality demands (Ramler and Wolfmaier, 2006).<br />

Shahamiri (2009) and Chin-Yu (2005) states that improvement of software quality and reliability through<br />

defect detection is a goal of software testing as is risk mitigation (Frank and Weyuker, 2000).<br />

Software testing drives are commonly beleaguered by constraints such as time, cost, and skill. These<br />

constraints risk the realisation of software testing goals and reduce test effectiveness. Understanding<br />

how to mitigate this risk is a key-factor for achieving successful software testing. Therefore the research<br />

question asked here is<br />

What test approaches are most likely to reduce the negative impact of test constraints on test<br />

effectiveness?<br />

Hence this research aims to identify software test tools or approaches likely to increase test effectiveness<br />

in the presence of test constraints.<br />

450


2. Literature review<br />

2.1 Definitions and terminology<br />

Grafton Whyte and Donovan Lindsay Mulder<br />

Software testing is defined as the observation of the execution of software based systems in order to<br />

verify that the system behaves as expected and to identify defects in the system under test (Bertolino,<br />

2007).<br />

Software testing effectiveness is defined as the number of defects found through software testing<br />

divided by the total number of defects (Vallespir and Herbert, 2009).<br />

For the purposes of this paper the researcher defines a:<br />

Test tool as any methodology, process or know-how that contributes to software testing and as such<br />

is not limited to software based test tools.<br />

Test approach as one or a combination of test tools used to implement software testing.<br />

Test constraint as any factor that inhibits the software testing process from achieving the desired<br />

levels of performance.<br />

Test Design refers to test case content and test case size.<br />

2.2 Test constraints<br />

According to a study by Chin-Yu (2005), defect detection rate is impacted by the skill level of test<br />

personnel and size of the project. As does testability (Berner, et al., 2005) of the system. Mirarab and<br />

Tahvildari (2008) argue that time constraints impact upon regression testing and negatively impact on the<br />

cost effectiveness of test case prioritization methodologies. Bryce and Colbourn (2005) state that due to<br />

time and cost constraints entire test suites in most instances are not run.<br />

2.3 Test design<br />

Rothermel, et al. (2004) discusses four regression test methodologies from a test design perspective.<br />

Two design factors are considered: test case size and test case content. The four methodologies are<br />

“retest-all, regression test selection, test case prioritisation, and test suite reduction”. All four are shown to<br />

have increased cost effectiveness, defect detection capability with reduced test execution time with the<br />

application of relevant test case design. It is shown that large test cases increases the cost effectiveness<br />

and reduce test execution time in the retest-all, regression test selection, and test case prioritisation<br />

methodologies. This was not shown to be true for test suite reduction. It was shown that functional<br />

grouping within test suite reduction could lead to greater defect detection capability. This however might<br />

not be true for difficult-to-find defects in all four methodologies. It is also shown that large test cases have<br />

a greater defect detection capability when compared to smaller test cases. Smaller test cases are more<br />

prone to finding hard to detect defects and lend itself to more effective test case prioritisation and<br />

regression test selection.<br />

2.4 Test selection methods<br />

One of the most common test selection methodologies is test case prioritisation. Test cases are ordered<br />

according to explicit test selection criteria. This is to enhance the defect detection rate early in the test<br />

process (Rothermel, et al., 2004; Srikanth and Williams, 2005) and decrease test execution time<br />

(Sampath, et al., 2008) with respect to the realisation of specific testing goals such as test all high risk<br />

functionality. TSC could be code coverage, possibility of defect existence and defect detection potential<br />

(Sampath, et al., 2008; Do, et al., 2008). Other TSC used to detect high risk defects quickly are<br />

“requirement volatility, customer priority, implementation complexity and fault proneness” (Srikanth and<br />

Williams, 2005).<br />

Regression test selection increases cost effectiveness however there are trade-offs one of which is<br />

between safety and efficiency. Regression test selection methods assume that test cases that do not test<br />

changed functionality will not detect defects (Engström 2010). The trade off here isthe risk of not finding<br />

all defects versus cost effectiveness and short test execution cycles.<br />

Another common test selection methodology is test suite reduction. The goal of test suite reduction is to<br />

reduce the cost of regression testing (Parse, et al., 2009) by satisfying all test requirements with the least<br />

amount of test cases (Zhang, et al., 2008a, b). Another aim of test suite reduction is to permanently<br />

451


Grafton Whyte and Donovan Lindsay Mulder<br />

remove redundant test cases from a test suite. This leads to reduced cost and time of test execution<br />

verification and test suite management (Rothermel, et al., 2004). Code and functional coverage and<br />

defect detection potential are commonly used as test selection (reduction) criteria (Parse and Khalilian<br />

and Fazlalizadeh, 2009). Discarding test cases can quickly result in a significant decrease in the defect<br />

detection capability of reduced test suites.<br />

2.5 Test oracles<br />

Test oracles are defined as an accepted dependable source of specified input and expected output of<br />

software behaviour and a means of reconciling expected and actual behaviour (Shahamiri, et al., 2009).<br />

From the definition it is clear that test oracles have a significant impact on the defect detection capability<br />

and cost effectiveness of a test suite which is a view supported by Memon, et al. (2003) as well as<br />

Memon and Xie (2004, 2005). A strong test oracle will have many verification points thus increasing test<br />

development, execution and verification time but increasing defect detection capability. A weak test<br />

oracle could result in reduced test execution time though this might be due to misleading or incomplete<br />

oracle information. In this case there are risks of defects not being detected. Memon and Qing (2005)<br />

also conclude that test cases lose their defect detection capability substantially, through decrepit test<br />

oracles; comprehensive test oracles employed at the end of the execution of a test case yields the best<br />

cost benefit ratio; some test oracles only detect defects during a small window of opportunity and<br />

comprehensive and regular application of comprehensive test oracles can counteract the impact of small<br />

test cases. This is supported by Rothermel, et al. (2004) who also states that large test cases minimises<br />

the impact of weak test oracles on test effectiveness.<br />

2.6 Smoke testing<br />

Smoke testing is used to detect defects early in the software development lifecycle (Memon, et al. 2003).<br />

It is widely accepted that smoke testing leads to lower defect resolution costs, lower costs of formal<br />

testing, enhanced software quality and risk minimization.Memon and Xie (2004) conclude that smoke<br />

tests detect greater than 60% of defects; small smoke tests execute a large percentage of code; larger<br />

smoke tests are able to detect more defects; test oracles have a substantial impact upon the defect<br />

detection capability of smoke tests.<br />

2.7 Automated testing<br />

Shahamiri and Kadir and Mohd-Hashim (2009) argue that test automation has been the one method used<br />

to decrease the costs of software testing. This is supported by Ramler and Wolfmaier (2006) and Karhu,<br />

et al. (2009) who also states that automated testing can be used in place of manual testing when time is<br />

a constraint. This is further supported by Zhu, et al. (2006) where it is stated that in order to reduce the<br />

costs and improve software test effectiveness it is critical to automate the testing process.<br />

2.8 Literature review conclusion<br />

The Software Test Tool Model (figure 1) serves as a summary of the literature review. This model<br />

summarises the relationship between Test Tools (Test Selection Criteria, Test Selection Methods, and<br />

Test Execution Methods) and Test Effectiveness as conceptualised for the purpose of this research.<br />

Derived from the ideas and concepts discussed in the literature review, this model provides a visual map<br />

of how all the test tools and test selection criteria fit together to form a software testing approach and how<br />

these factors collectively can increase test effectiveness.<br />

The Software Test Tool model works as follows: the Test Selection Criteria impacts upon the<br />

effectiveness of the Test Selection Methods which in turn impacts Software Test Effectiveness. Test<br />

Selection Criteria applied to Test Cases which in turn impacts upon Software Test Effectiveness. Test<br />

Design and Test Oracles impacts upon the defect detection rate and capability of Test Cases which in<br />

turn impacts upon Software Test Effectiveness. Test Selection Criteria impacts the test execution time of<br />

Test Execution Methods which in turn impacts upon Software Test Effectiveness.<br />

From the literature it is concluded that:<br />

Skill of human resources, project size, inadequate requirements, software testability, time, cost and<br />

test design are test constraints falling into the categories of time, cost and skill.<br />

Retest-all, regression test selection, test case prioritisation, test suite reduction, smoke testing, test<br />

automation, test oracles and test design are determined from the literature to be test tools used to<br />

minimise the impact test constraints on test effectiveness.<br />

452


Grafton Whyte and Donovan Lindsay Mulder<br />

Test selection criteria such as code coverage, possibility of defect existence, defect detection<br />

potential, test case design (test case content and size);and requirement volatility, customer priority,<br />

implementation complexity and fault proneness are identified as test selection criteria which can<br />

support test case prioritisation.<br />

These test tools and test selection criteria were found to directly affect test effectiveness through four<br />

aspects which featured prominently in the reviewed literature; these are (a) defect detection<br />

capability, (b) defect detection rate, (c) cost effectiveness and (d) test execution time.<br />

Figure 1: Software effectiveness model<br />

3. Research methodology<br />

The ideas discussed in the literature review embody the Goal-Question Metric Approach template<br />

proposed by Basili et al. (1994): They propose analysing software testing tools, software test selection<br />

criteria and software test constraints for the purpose of knowing which software testing tools and test<br />

case selection criteria or a combination thereof increases software test effectiveness in the presence of<br />

software test constraints with respect to their usage and possible software testing improvement potential<br />

from the view point of software testing experts and professionals.<br />

There are a plethora of test tools and test selection criteria in the current literature that can arguably<br />

minimise the negative impact of test constraints on test effectiveness. The usage of these test tools and<br />

test selection criteria differ widely, while information about the actual usage in the presence of test<br />

constraints in project situations seems to be missing. To assess the extent to which test tools minimise<br />

the negative impact, a research instrument was needed that could be measure the perceived impact of<br />

test tools on test effectiveness in the presence of test constraints. No such instrument was found in the<br />

literature. Most if not all of the research in the literature dealt with test tools used under experimental<br />

conditions and not in real projects. Also it is not certain how these test tools and test selection criteria<br />

would contribute to the overall test effectiveness in live projects. Therefore a research instrument was<br />

developed based on test tools and test selection criteria found in the literature and that could be used to<br />

measure the perceived impact on actual projects.<br />

453


Grafton Whyte and Donovan Lindsay Mulder<br />

A survey questionnaire consisting of closed questions was used as the research instrument. Quantitative<br />

data was collected. The research constructs are test tools, test case selection criteria, test constraints<br />

and test effectiveness. The survey was designed to elicit data from practitioners’ perspectives and<br />

experience. The aim was to identify the perceived impact (a) test tools have on test effectiveness, (b) test<br />

selection criteria on test effectiveness, (C) test constraints on test effectiveness and (d) actual test<br />

effectiveness of testing activities. The research instrument was piloted using 2 case studies. The pilot<br />

study was interview based in order to gauge first-hand the effectiveness of the survey instrument. The<br />

instrument yielded coherent results with minor modifications.<br />

The research population was the world of software testing professionals in which one of the authors was<br />

a recognised member. Convenience sampling was used to derive the sample from the population. Each<br />

respondent was asked to provide data for three projects. Each project was treated as a single case. In<br />

total 43 cases was reported on by the respondents. The survey questionnaire was sent to 14 members of<br />

the sample of which 2 did not respond effectively achieving a response rate 85.7% with respect to the<br />

initial sample. However the sample size grew to 18 with a total of 16 responses (88.9%). Of the possible<br />

47 possible cases 43 were deemed acceptable as the value of software test effectiveness was omitted in<br />

4 cases. Respondents were located in geographically dispersed locations (USA, England, Ireland,<br />

France, South Africa, India and Australia). The predominant roles were test managers and test leads (see<br />

figure 2). The case studies covered a range of industries with financial services and telecoms dominating<br />

(see figure 3).<br />

3.1 Validity<br />

Construct Validity: Observations were based on the respondents experience and as such might have led<br />

to some differences in interpretation to mitigate, definitions of all test variables were defined in the<br />

questionnaire.<br />

External Validity: 43 case studies were included in this research. The results can be generalised to a<br />

certain extent as the case studies are from different industries and from geographically dispersed<br />

locations .Although convenience sampling was used to select respondents due to the dispersed nature of<br />

the sample it can be considered a fair representation of the population.<br />

Internal Validity: The survey questionnaire was designed with a certain goal in mind; there were no<br />

extraneous questions therefore, it is reasonable to conclude that the instrument had internal validity.<br />

3.2 Reliability<br />

Joppe (2000) defines reliability as: “The extent to which results are consistent over time and an accurate<br />

representation of the total population under study is referred to as reliability and if the results of a study<br />

can be reproduced under a similar methodology, then the research instrument is considered to be<br />

reliable.” The survey tool is based upon the Goal-Question Metric Approach template proposed by Basili,<br />

et al. (1994), the tool demonstrated acceptable levels of reliability in their research therefore, reliability is<br />

assumed.<br />

Figure 2: Distribution of roles<br />

454


Figure 3: Distribution of industries<br />

4. Research results<br />

Grafton Whyte and Donovan Lindsay Mulder<br />

According to the Goal-Question Metric approach (Basili et al.1994) for measurement to be useful from an<br />

organisational view point it must be goal driven and specific. The purpose of measurement in this<br />

research is to identify the testing tools most likely to overcome organisational constraints and yield<br />

optimal test effectiveness.<br />

From the outset two goals were set for this research:<br />

1. Identify individual Test Tools (TT) and Test Selection Criteria (TSC) most likely to yield best Test<br />

Effectiveness (TE). It was determined that the best statistical approach for identifying these<br />

constructs would be correlation analysis.<br />

2. Identify the combination of Test Tools (TT) and Test Selection Criteria (TSC) most likely to yield<br />

best Test Effectiveness (TE). It was determined that the best statistical approach for identifying these<br />

constructs would be multiple regression analysis.<br />

4.1 Descriptives<br />

Constructs for this research were organised into three construct groups:<br />

Software Test Tools (TT) – Table 1.0lists each tool as presented to respondents, a definition was also<br />

added to ensure a common understanding. Tools are listed here with group descriptions where<br />

appropriate. Respondents were asked to rate for each project whether the test tool increased test<br />

effectiveness. Their responses were graded on a Likert scale of 1 to 5 denoting ‘complete disagreement’<br />

to ‘complete agreement’, respectively. Sample size (N=43) in every case.<br />

Table 1: Software Test Tools<br />

Type Description<br />

Test Design<br />

Variable<br />

Label<br />

Mean<br />

Std. Dev.<br />

Test case content TT1 3.77 1.269<br />

Test case granularity TT2 3.33 1.375<br />

Test oracles<br />

Test selection method<br />

TT3 2.63 1.865<br />

Regression test selection TT4 3.35 1.526<br />

455


Grafton Whyte and Donovan Lindsay Mulder<br />

Type Description Variable<br />

Label<br />

Mean<br />

Std. Dev.<br />

Re-test all TT5 2.65 1.462<br />

Test suite reduction TT6 2.19 1.577<br />

Test case prioritisation TT7 3.91 1.151<br />

Test execution method<br />

Automated testing TT8 2.70 1.859<br />

Manual testing TT9 4.19 1.220<br />

Smoke testing TT10 3.60 1.788<br />

Software Test Selection Criteria (TSC) – Table 2.0 list each test selection criteria. Respondents<br />

received them with definitions and were asked to rate projects on each criterion; whether these test<br />

selection criteria increase software test effectiveness. Responses were graded on a Likert scale of 1 to 5<br />

denoting ‘no increase’ to ‘significantly increased’, respectively.<br />

Table 2: Software Test Selection Criteria<br />

Type Description Variable<br />

Mean<br />

Std. Dev.<br />

Label<br />

Test Case<br />

Code coverage TSC1 1.40 1.400<br />

Functional coverage TSC2 3.81 1.118<br />

Defect detection capability TSC3 3.09 1.702<br />

Setup time TSC4 2.74 1.706<br />

Implementation complexity TSC5 2.86 1.207<br />

Execution time TSC6 3.58 1.239<br />

Requirements<br />

Customer priority TSC7 3.63 1.543<br />

Implementation complexity TSC8 2.88 1.562<br />

Fault proneness TSC9 2.51 1.907<br />

Volatility TSC10 2.79 1.767<br />

Test Case Composition<br />

Test case content TSC11 3.53 1.182<br />

Test case granularity TSC12 2.88 1.219<br />

No test selection criteria TSC13 1.74 1.853<br />

Test Constraints (TC) – Table 3 list the main test constraints encountered by test projects, again<br />

respondents were asked to rate projects on each criterion, whether these test constraints hindered<br />

software test effectiveness. Responses were graded on a Likert scale of 1 to 5 denoting ‘no impact’ to<br />

‘significantly impacted’, respectively.<br />

Table 3: Test Constraints<br />

Type Description<br />

Test Constraints<br />

Variable<br />

Label<br />

Mean<br />

Std. Dev.<br />

Time Time 3.72 1.260<br />

Cost Costs 2.93 1.534<br />

Skills Skills 2.81 1.332<br />

The relationship between these variables and test effectiveness can be represented in the following linear<br />

function<br />

Test effectiveness (TE) = Test Tools (TT) + Test Selection Criteria (TSC) – Test Constraints (TC)<br />

Subsequent correlation and regression tests explored the applicability of this function.<br />

4.2 Goal 1 - correlation tests<br />

Pearson correlation was run for each of the constructs independently (assuming two of the three<br />

constructs were zero) to see which of the underlying variables correlates with the construct Test<br />

Effectiveness (TE).<br />

456


Grafton Whyte and Donovan Lindsay Mulder<br />

The results from Table 4 suggest that regression testing selection (TT4) and smoke testing (TT10)<br />

correlate strongly with test effectiveness (TE), with 54% and 44% of the variability in TE explained by<br />

these two variables, with both achieving significance levels in the 99 percentile, suggesting these results<br />

are very reliable. Two further variables, test case prioritisation (TT7) and automated testing (TT8)<br />

correlate less strongly with TE (35% and 38%, respectively), but still achieve significance levels in the 95<br />

percentile.<br />

Table 4: Test Tools (TT) correlated with TE<br />

Type Description Variable<br />

Pearson<br />

Significance<br />

Test Design<br />

Label<br />

Correlation<br />

(2-tailed)<br />

Test case content TT1 -0.036 0.821<br />

Test case granularity TT2 0.099 0.527<br />

Test oracles<br />

Test selection method<br />

TT3 0.206 0.185<br />

Regression test selection TT4 0.536 0.000**<br />

Re-test all TT5 0.016 0.919<br />

Test suite reduction TT6 0.106 0.501<br />

Test case prioritisation<br />

Test execution method<br />

TT7 0.347 0.023*<br />

Automated testing TT8 0.384 0.011*<br />

Manual testing TT9 -0.129 0.409<br />

Smoke testing TT10 0.444 0.003**<br />

From Table 5 we note that one variable, defect detection capability (TSC3), achieved correlation of 39%<br />

at significance levels in the 99 percentile (0.010). Other variables that achieved high significance levels in<br />

the 95 percentile are functional coverage, customer priority, test case content and no test selection<br />

criteria (test all) with respective correlations of 31%, 35%, 31% and 34%.<br />

Table 5: Test Selection Criteria (TSC) correlated with TE<br />

Type Description Variable<br />

Pearson<br />

Significance<br />

Test Case<br />

Label<br />

Correlation<br />

(2-tailed)<br />

Code coverage TSC1 0.219 0.159<br />

Functional coverage TSC2 0.315 0.039*<br />

Defect detection capability TSC3 0.390 0.010**<br />

Setup time TSC4 0.270 0.080<br />

Implementation complexity TSC5 0.196 0.207<br />

Execution time<br />

Requirements<br />

TSC6 0.071 0.649<br />

Customer priority TSC7 0.354 0.020*<br />

Implementation complexity TSC8 0.017 0.914<br />

Fault proneness TSC9 0.155 0.320<br />

Volatility<br />

Test Case Composition<br />

TSC10 0.106 0.499<br />

Test case content TSC11 0.306 0.046*<br />

Test case granularity TSC12 0.167 0.285<br />

No test selection criteria TSC13 0.343 0.024*<br />

There were no significant correlations with any variables in the test constraint construct (TC).<br />

4.3 Goal 2 – multiple regression testing<br />

Having identified variables to emerge from independent tests of association with Test Effectiveness,<br />

attention was turned to testing the linear function in totality using multiple regression analysis to see if the<br />

variables that emerged in the correlation test would hold or would new variables emerge as predictors of<br />

test effectiveness.<br />

457


Grafton Whyte and Donovan Lindsay Mulder<br />

Data for the model were analysed using linear multiple regression analysis. The procedure estimates the<br />

coefficients (beta) of one or more independent variables to predict (R²) the value of a single dependent<br />

variable. Variables are systematically entered and removed from the equation using the stepwise method<br />

to determine the line of best fit.<br />

Table 6: Linear Regression model<br />

Dependent Variable Independent Variables Beta R Adjusted R² Significance<br />

Test Effectiveness<br />

(TE)<br />

0.718 0.479 .000<br />

Regression test selection<br />

(TT4)<br />

0.465 .000<br />

Smoke testing (TT10) 0.340 .005<br />

No test selection criteria<br />

(TSC13)<br />

0.337 .004<br />

The model suggests that 48% of the variability in the construct Test Effectiveness (TE) is explained by<br />

the variables TT4, TT10 and TSC13. The proportion that each of the independent variables explain is<br />

indicated by their beta percentages, 46%, 34% & 33% respectively.<br />

5. Discussion<br />

Test tools and test selection are applied quite extensively in the case studies investigated. Given the<br />

diversity of the population it can be assumed this is likely to be the case for the Software Testing industry<br />

globally. This indicates that software testing practice is becoming more rigorous and formalised and to a<br />

certain extent scientific, which is a good sign for the software development industry and business as a<br />

whole.<br />

The results suggest regression test selection (TT4) and smoke testing (TT10) significantly, and to a<br />

lesser extent test case prioritisation (TT7) and automated testing TT11) correlated with Test<br />

Effectiveness. These Test Tools can each be applied individually to testing activities in order to increase<br />

Test Effectiveness.<br />

In practice these Test Tools are most associated with increased Test Effectiveness and their use is more<br />

likely to increase the probability that a software development project will meet its strategic goals.<br />

Smoke testing is used to detect defects before more expensive formal testing. Automated testing<br />

significantly reduces test execution time and also increases functional coverage during testing in shorter<br />

test cycles. Individually the application of these Test Tools will lead to early detection of defects, reduced<br />

cost of testing, quicker time to market and increased software quality.<br />

Regression test selection and test case prioritisation are dependent upon adequate test selection criteria.<br />

One test selection criterion that correlated strongly with Test Effectiveness is defect detection capability<br />

(TSC3). Other test selection criteria that correlated with slightly less significance are functional coverage<br />

(TSC2), customer priority (TSC7), test case content (TSC11) and no test selection criteria (TSC13).<br />

Defect detection capability (TSC3) is an imperative design consideration. It is the make or break factor in<br />

the realisation of software quality through increased test effectiveness. It plays a pivotal role in reducing<br />

the risk of software going to market with undetected defects. Maximised functional coverage (TSC2)<br />

during test execution reduces the risk of parts of a software system being untested after it has been<br />

released to market. Test case content (TSC11) directly relates to functional coverage and as such is<br />

used to measure functional coverage of software testing, thus giving the business a risk based view that<br />

can be used to determine the readiness of a software system before release to market. Customer priority<br />

(TSC7) allows software development businesses to quickly meet the most important goals of the<br />

customer. No test selection criteria (TSC13) translates into testing of the entire software system thereby<br />

mitigating all risk posed by inadequate functional coverage. However this is an expensive approach to<br />

software testing and requires no or very little time constraints.<br />

Individually these Test Selection Criteria increase software quality, customer satisfaction thereby possibly<br />

enhancing and preserving a business’s good reputation.<br />

458


Grafton Whyte and Donovan Lindsay Mulder<br />

Applying multiple regression analysis to the combined Test Tools (TT) and Test Selection Criteria (TSC)<br />

identified the combined variables of regression test selection (TT4), smoke testing (TT10) and no test<br />

selection criteria (TSC13) to have the most significant impact upon Test Effectiveness.<br />

In practice this result suggest that smoke testing (TT10) should be used for early detection of major<br />

defects after which regression test selection (TT4) should be applied to the software under test and that<br />

no test selection (TSC13) be applied to test cases, therefore test everything. Testing everything is an<br />

approach best used for major software release prior to go-to-market. Though this is expensive it mitigates<br />

the risk of defects ‘making it out into the wild’ resulting in damaged reputation and possible significant<br />

economic loss and even loss of life. During minor releases where it does not make economic sense to do<br />

exhaustive testing with full functional coverage one or more of the Test Selection Criteria identified<br />

through the correlation analysis should be applied. This will lead to reduced test execution costs and time<br />

with increased defect detection rate in system under test.<br />

The data analysis strongly suggested a saturation point in the application of the number of Test Tools<br />

and Test Selection Criteria, exactly at which point this achieved is debatable but initial indications from<br />

this research suggests after a 80% level of Test Effectiveness has been achieved. The total number of<br />

Test Tools and Test Selection Criteria in this study is twenty-three (23). Correlation and multiple<br />

regression analyses reduced this to nine and three respectively, suggesting a high-degree of over-lap<br />

between tests. In business terms it would seem not to make sense to try and achieve levels of Test<br />

Effectiveness beyond the 80 percentile point.<br />

Interestingly, neither correlation nor multiple regression analyses revealed a significant relationship<br />

between Test Constraints and Test Effectiveness. Given that organisational constraints are always<br />

uppermost in practitioner’s minds. One reason for this absence could be due to the fact that the sample<br />

was made up of seasoned test experts, who through their experience have learnt how to mitigate the<br />

impact of Test Constraints on Test Effectiveness.<br />

6. Conclusion<br />

The aim of this paper was to identify from the plethora of Test Tools used in practice to determine<br />

1. Which individual test tools are most likely to yield optimal test effectiveness and,<br />

2. Which combination of test tools is most likely to yield optimal test effectiveness and mitigate the<br />

effect of test constraints<br />

Data from forty three cases across various industries and countries were collected to identify the current<br />

application of Test Tools in practise. The research was designed using the Goal-Question Metric<br />

approach and the data analysed using correlation analysis to identify individual Test Tools and Test<br />

Selection Criteria most closely associated with Test Effectiveness and; multiple regression analysis to<br />

identify the combination of Test Tools and Test Selection Criteria that would lead to optimum Test<br />

Effectiveness.<br />

The correlation analysis identified nine variables; Four Test Tools (TT): regression test selection (TT4),<br />

smoke testing (TT10), test case prioritisation (TT7) and automated testing (TT11) and, five Test Selection<br />

Criteria (TSC): defect detection capability (TSC3), functional coverage (TSC2), customer priority (TSC7),<br />

test case content (TSC11) and no test selection criteria (TSC13), allowing test practitioners to pick and<br />

mix the various approaches given the specific constraints.<br />

The multiple regression analysis identified the combined variables of regression test selection (TT4),<br />

smoke testing (TT10) and no test selection criteria (TSC13) to have the most significant impact upon Test<br />

Effectiveness.<br />

The main value of this research is that it begins to spell out for the growing community of test<br />

practitioners who have limited experience in the field of software testing and are confronted by a<br />

bewildering array of test options, which individual and combined test techniques will most likely help them<br />

achieve optimal test effectiveness.<br />

Areas of further research:<br />

A saturation point in the application of Test Tools and Test selection criteria is alluded to in this<br />

paper, more evidence to support this idea is required and to pinpoint the threshold. This will enable<br />

459


Grafton Whyte and Donovan Lindsay Mulder<br />

Of the research sample the majority of test cases were overwhelmingly successful, further research<br />

needs to be conducted on cases which are not successful to ascertain if the emergent variables<br />

remain consistent.<br />

References<br />

Basili, V.R. Caldiera, G., Rombach, H.D. 1994, 'The Goal Question Metric Approach', Encyclopedia of Software<br />

Engineering<br />

Berner, S., Weber, R., Keller, R.K. 2005, 'Observations and Lessons Learned from Automated Testing', Zühlke<br />

Engineering AG Zürich-Schlieren Switzerland, 2005.<br />

Bertolino, A 2007, 'Software Testing Research: Achievements, Challenges, Dreams', no. Future of Software<br />

Engineering (FOSE'07).<br />

Bryce, R.C., Colbourn, C.J. 2005, 'Test Prioritization for Pairwise Interaction Coverage', A-MOST’05, Copyright 2005<br />

ACM 1-59593-115-5/00/0004, St. Louis, Missouri, USA.<br />

Chin-Yu, H. 2005, 'Performance analysis of software reliability growth models with testing-effort and change-point',<br />

The Journal of Systems and Software, vol 76, pp 181 - 194.<br />

Do, H., Mirarab, S., Tahvildari, L. 2008, 'An Empirical Study of the Effect of Time Constraints on the Cost-Benefits of<br />

Regression Testing', SIGSOFT 2008/FSE-16, November 9–15, Copyright 2008 ACM 978-1-59593-995-1,<br />

Atlanta, Georgia, USA.<br />

Engel, A., Last, M. 2007, 'Modeling software testing costs and risks using fuzzy logic paradigm', The Journal of<br />

Systems and Software, 2007.<br />

Engström, E 2010, 'Regression Test Selection and Product Line System Testing', 2010 Third International<br />

<strong>Conference</strong> on Software Testing, Verification and Validation, IEEE Computer Society, Paris.<br />

Frank, P. G. , Weyuker, E. J. 2000, 'Testing software to detect and reduce risk', The Journal of Systems and<br />

Software, 2000.<br />

Joppe, M. 2000. 'The Research Process', Available at [Accessed 08March<br />

2009]<br />

Karhu, K., Repo, T., Taipale, O., Smolander, K. 2009, 'Empirical Observations on Software Testing Automation',<br />

2009 International <strong>Conference</strong> on Software Testing Verification and Validation, IEEE Computer Society, Denver<br />

Colorado.<br />

Memon, A.M., Bananerjee, I., Hashmi, N., Nagarajan, A. 2003, 'DART: A Framework for Regression Testing<br />

Nightly/daily Builds of GUI Applications', Proceedings of the International <strong>Conference</strong> on Software Maintenance<br />

(ICSM’03), IEEE Computer Society.<br />

Memon, A.M., Xie, Q. 2004, 'Empirical evaluation of the fault-detection effectiveness of smoke regression test cases<br />

for GUI-based software', Proceedings of the 20th International <strong>Conference</strong> on software Maintenance (ICSM<br />

2004), Chicago, USA.<br />

Memon, A.M., Xie., Q. 2005, 'Studying the Fault-Detection Effectiveness of GUI Test Cases for Rapidly Evolving<br />

Software', IEEE Transactions on Software Engineering, 2005.<br />

Parse, S., Khalilian, A., Fazlalizadeh, Y. 2009, 'A New Algorithm to Test Suite Reduction Based on Cluster Analysis',<br />

2009.<br />

Ramler, R., Wolfmaier, K. 2006, 'Economic Perspectives in Test Automation: Balancing Automated and Manual<br />

Testing with Opportunity Cost"'.<br />

Rothermel, G., Elbaum, S., Malishevsky, A.G. , Kallakuri, P., Xuemei Q. 2004, 'On Test Suite Composition and Cost-<br />

Effective Regression Testing', ACM Transactions on Software Engineering and Methodology, 2004.<br />

Sampath, S., Ren´ee C. B., Gokulanand, V., Vani K., A. Gunes¸ K. 2008, 'Prioritizing User-session-based Test Cases<br />

for Web Applications Testing', International <strong>Conference</strong> on Software Testing, Verification, and Validation, IEEE<br />

Computer Society.<br />

Shahamiri, S.R., Kadir, W.M.N.W, Mohd-Hashim. S.Z. 2009, 'A Comparative Study on Automated Software Test<br />

Oracle Methods', Fourth International <strong>Conference</strong> on Software Engineering Advances, IEEE Computer Society.<br />

Srikanth, H., Williams, L. 2005, 'On the Economics of Requirements-Based Test Case Prioritization', EDSER'05,<br />

Copyright 2005 ACM 1-59593-118-X/05/0005, St. Louis, Missouri, USA.<br />

Srikanth, H., Williams, L., Osborne, J. 2005, 'System Test Case Prioritization of New and Regression Test Cases',<br />

International Symposium on Empirical Software Engineering, IEEE Computer Society.<br />

Vallespir, D., Herbert, J. 2009, 'Effectiveness and Cost of Verification Techniques', IEEE Computer Society, Mexican<br />

International <strong>Conference</strong> on Computer Science.<br />

Zhang, R., Jiang, J., Yin, J., Jin, A., Lou, J., Wu, Y. 2008, 'A New Method for Test Suite Reduction', The 9th<br />

International <strong>Conference</strong> for Young Computer Scientists, IEEE Computer Society.<br />

Zhang, X., Xu, B., Chen, Z., Nie. C., Li, L. 2008, 'An Empirical Evaluation of Test Suite Reduction for Boolean<br />

Specification-based Testing', The Eighth International <strong>Conference</strong> on Quality Software, IEEE Computer Society.<br />

Zhu, H., Horgan, J.R., Cheung, S.C., Li, J.J. 2006, 'The First International Workshop on Automation of Software<br />

Test', ICSE’06, ACM 1-59593-085-X/06/0005, Shanghai, China.<br />

460


Strategic Considerations for Effective Mapping of<br />

Educational Technology to <strong>Academic</strong> Outcomes<br />

Hossein Zadeh, Arthur Shelley and Rod McCrohan<br />

RMIT University, Melbourne, Australia<br />

hossein.zadeh@rmit.edu.au<br />

arthur.shelley@rmit.edu.au<br />

rod.mccrohan@rmit.edu.au<br />

Abstract: Advancements in IT and web-based technologies in the past two decades have the potential to<br />

revolutionize higher education, but realizing this potential seems to remain an elusive goal. Educators in diverse<br />

contexts are constantly exploring innovative technologies that claim to add value to traditional classroom knowledge<br />

delivery and impact the course delivery and design in colleges and universities. Even though the use of technology<br />

in higher education has substantially increased over the past few years, this use has been primarily limited to content<br />

management and delivery. There are, however, much broader strategic issues around alignment between the suite<br />

of learning technology tools promoted by universities and the academic plan, learning and teaching priorities, and<br />

expected graduate capabilities. To date many universities have not undertaken a full strategic assessment of the<br />

issues and implications of the introduction and sustaining of educational technologies and how these can best be<br />

utilized to fulfill academic policy and add value to student learning experiences and outcomes. We conducted a focus<br />

group of a number of innovative and early technology-adopter university teachers in which they discussed how<br />

educational technology tools (Ed-Tools) can be embedded into courses such that they develop students’ analytical,<br />

sense-making and synthesis skills whilst enhancing their decision-making and delivery capabilities. This study<br />

highlights the importance of effectively using educational technology to support academic staff with their teaching<br />

and learning. A distinguishing factor of this study is that it looks at the issue from a holistic point of view, and not<br />

merely from the teachers’, or the administration, or the students’ point of view. We have compiled a list of barriers<br />

and also a list of recommendations. Findings of this exploratory study are presented intentionally in the form of a<br />

discussion paper rather than a traditional academic paper. It is hoped this presentation format opens up the issues<br />

for in-depth consideration by learning and teaching policy makers at universities and colleges.<br />

Keywords: educational technology, ed-tech, ed-tools, teaching and learning, graduate capabilities<br />

1. Introduction<br />

Advancements in IT and web-based technologies in the past two decades have the potential to<br />

revolutionize higher education, but realizing this potential seems to remain an elusive goal. Educators in<br />

diverse contexts are constantly exploring innovative technologies that claim to add value to traditional<br />

classroom knowledge delivery and impact the course delivery and design in colleges and universities<br />

(Barnett, Keating, Harwook, & Saam, 2004 as cited in Ajjan & Hartshorne, 2008). Numerous past studies<br />

have shown that technology use in higher education has increased (Maloney, 2007). However; this use<br />

has been primarily limited to content management and delivery, such as accessing course materials.<br />

Dalsgaard (2006) noted that whether the focus is on distance education or campus based education,<br />

universities all over the world are using information management systems to support and improve<br />

learning within their institutions. Learning, however, is far more than transfer of content. In the<br />

experience of the authors, the extent to which these technologies are actually applied varies to a great<br />

extent; some courses use technology merely as content management while others engage in a variety of<br />

more innovative approaches to actively engage students in interactive learning experiences.<br />

To be effective, learning and teaching requires preparing the students to cope with making decisions in<br />

an unpredictable future environment. The Web, for example, has grown into a global information space<br />

with more than a billion users and has entered a new, more social and participatory phase where people<br />

are engaging in different types of online collaborative activities (Anderson, 2007). This new form of the<br />

Web, often referred to as Web 2.0, allows people to communicate knowledge, share resources and<br />

participate in social networks (Ding, Jacob, Caverlee, Fried & Zhang, 2009). With the emergence of Web<br />

2.0, students no longer access the Web only for course information; instead they access and create<br />

collective knowledge through social interactions (Maloney, 2007).<br />

1.1 Learning support systems<br />

Almost all digital object support systems (online or local) share many of the same functions, but differ in<br />

the environments in which they are used. Course management systems (CMSs), learning content<br />

management systems (LCMSs) and virtual learning environments (VLEs) are generally used to facilitate<br />

461


learning in an academic setting. While both the terms CMS and VLE can be used interchangeably, these<br />

terms seem to be region specific. CMS is more commonly used in North America and VLE is more<br />

prevalent in <strong>European</strong> countries (Weller 2007 as cited in Daniels, 2009). These eLearning acronyms are<br />

further complicated with the more recent introduction of the PLE or personal learning environment,<br />

intended to shift greater control of learning environments to the learners by giving them greater flexibility<br />

in choosing their learning tools (Sclater 2008 as cited in Daniels, 2009).<br />

Learning management systems (LMS) typically are used to train employees in business environments.<br />

CMSs are widely used by colleges and universities to manage and deliver course content presented via<br />

the web. For example, MIT uses Stellar; the University of Michigan uses CHEF; and Stanford uses<br />

CourseWork. Often these systems designed for a particular institution are not publicly available.<br />

Educational institutions without adequate resources to develop a personalized CMS can employ an open<br />

source CMS such as Moodle, LRN, or Sakai. For institutions needing additional technical support,<br />

proprietary learning systems are available from companies such as Blackboard and Desire2Learn. The<br />

primary functions of CMSs are to organize and distribute course content, administer learning exercises or<br />

quizzes, and track student progress. While CMS software is often used to manage distance learning<br />

courses, it is just as popular in supporting face-to-face instruction. It should be noted that all these<br />

information management systems are based on web 2.0 technologies (Daniels, 2009).<br />

1.2 Web 2.0<br />

Web 2.0 also referred to as ‘social software’ covers a range of tools that allows users to interact and<br />

share data with each other via the internet. Blogs, wikis, podcasts, social networking web sites (facebook,<br />

twitter, MySpace etc.) and social bookmarking sites, such as Delicious, and 3D environments such as<br />

Second Life are examples of some of the tools that are being used to share and collaborate in social,<br />

educational and business contexts. The Web 2.0 “read/write” idea is not new as there were listservs,<br />

groupware, and web-based communities linking people with common interests even before the advent of<br />

Web 2.0 (Alexander, 2006 as cited in Ajjan & Hartshorne, 2008).<br />

In the educational context Web 2.0 has reshaped both the way instructors teach and the way students<br />

learn. These new technologies have simplified content sharing and transformed knowledge creation,<br />

usage and distribution (Dearstyne, 2007). With the use of Web 2.0, students no longer access the web<br />

only for course information; instead they access and create collective knowledge through social<br />

interactions (Maloney, 2007 as cited in Ajjan & Hartshorne, 2008). Table 1 lists some of the more<br />

commonly used Web 2.0 educational technology tools in use today. Dalsgaard (2006) argues that Web<br />

2.0 technologies support self-governed, problem- based and collaborative learning processes unlike<br />

other educational viewpoints where the responsibility rests with the teacher to deliver knowledge while<br />

the learner passively receives it. Ferdig (2007, as cited in Ajjan & Hartshorne, 2008) also notes that Web<br />

2.0 applications can support pedagogical approaches such as active learning, social learning, and<br />

student publication, by providing environments and technologies that promote and foster these<br />

interactions.<br />

With the emergence of the new learning environment, it is imperative for universities to ensure that<br />

educational technologies have positive impact in assisting the current learning and teaching practices<br />

Table 1: List of commonly used Web 2.0 educational technology tools<br />

List of commonly used<br />

Ed-Tools<br />

Audio blogs, Video Blogs & Podcasts<br />

Wikis, Discussion boards, Instant<br />

Messaging (IM)<br />

E-portfolios & blogs<br />

Social bookmarks<br />

Tele and Video Conferencing<br />

Function<br />

Sharing ideas, information and content/creations<br />

Working and collaborating with others in a shared work area<br />

Collection and/or presentation of evidence of experiences, thinking over<br />

time, productions, etc.<br />

To exchange information, ideas, resources, materials<br />

Conversations for geographically spread resources<br />

462


Some examples of benefits of the use of Ed-Tools are presented in table 2<br />

Table 2: Examples of benefits from implementation of Ed-Tools<br />

Theme Comment Reference<br />

Interaction,<br />

communication and<br />

collaboration<br />

Knowledge creation<br />

Ease of use and flexibility<br />

Writing skills & motivation<br />

Insight<br />

A sense of community and<br />

resource sharing amongst users.<br />

Active and student-centered<br />

learning. Educators as facilitators<br />

User friendly and flexible learning<br />

technologies<br />

Peer assessment & feedback.<br />

Global exposure and writing<br />

proficiency<br />

Educators gain an insight in to the<br />

world of students<br />

Alexander, 2006; Brown & Adler, 2008;<br />

Hartshorne & Ajjan, 2009; Richardson, 2009;<br />

Dalsgaard, 2006<br />

Williams & An, 2010<br />

Alexander, 2006; Brown & Adler, 2008;<br />

Hartshorne & Ajjan, 2009; Richardson, 2009;<br />

Dalsgaard, 2006<br />

(Ajjan & Hartshorne, 2008)<br />

Williams & An, 2010<br />

These practices should support the seven principles of good practice in undergraduate education,<br />

namely:<br />

Good practice encourages interaction between students and faculty<br />

Good practice encourages interaction and collaboration between students<br />

Good practice uses active learning techniques<br />

Good practice gives prompt feedback<br />

Good practice emphasizes time on task<br />

Good practice communicates high expectations<br />

Good practice respects diversity – talents, experience, and ways of learning<br />

Furthermore, many universities consider as priority engagement with communities, enterprises and<br />

industry, and strive to produce graduates who are work-ready professionals.<br />

In order to prepare such graduates, universities need to have educational technologies adopted and<br />

embedded into course learning and teaching activities. So far the adoption does not appear to be<br />

universal across universities (or even within individual universities). In order to elicit and address the<br />

issues affecting the adoption and embedding of learning and teaching technologies, a group of<br />

academics who have already embedded a variety of learning and teaching technologies into their<br />

teaching practice were invited to participate in a focus group. A summary of the findings of the focus<br />

group is presented in this report.<br />

2. Process<br />

Leading educational technology adopters as well as key learning and teaching representatives from<br />

business schools in one of the largest universities in Australia were invited to a focus group. The focus<br />

group was facilitated by an external consultant specilising in Ed-Tools. The aim of the focus group was to<br />

identify key strategic themes and issues regarding the use and adoption of Ed-Tools in learning and<br />

teaching. Examples of barriers to adoption of Ed-Tools, as reported in the literature, were noted by the<br />

focus group (see Table 3). Specifically, the goal of the focus group was to address the question: How can<br />

educational technologies be adopted & embedded into course learning and teaching activities?<br />

The external facilitator guided the group through a series of questions designed to discover:<br />

How teaching & learning technologies are currently being used at the university;<br />

How industry use of ICT relates to student experience;<br />

The barriers to embedding teaching & learning technologies into the curriculum; and<br />

How the university could support staff in their adoption of teaching & learning technologies into the<br />

curriculum.<br />

463


Table 3: Examples of barriers to adoption of Ed-Tools<br />

<strong>Academic</strong>s<br />

Students<br />

Theme Comment Reference<br />

Technology adoption ‘one-at-a time’ approach to adoption and<br />

integration.<br />

McGee & Diaz, 2007<br />

Lack of integration with<br />

CMS<br />

Learners' changing<br />

expectations<br />

Institutional commitment<br />

and Support<br />

Uneasiness with nature of<br />

e-tools<br />

Technical issues<br />

Time and effort<br />

Emerging e-tools not integrated into CMS<br />

Changing student cohorts and expectations of<br />

instructional strategies<br />

Review and change/upgrade of CMS.<br />

Investment in E-tools<br />

Some students uncomfortable with openness<br />

of e-tools<br />

Technical glitches and insufficient technical<br />

support with e-tools<br />

E-tools take time away from learning subject<br />

matter content<br />

McGee & Diaz, 2007;<br />

Dalsgaard, 2006<br />

McGee & Diaz, 2007<br />

Williams & An, 2010<br />

Williams & An, 2010<br />

Williams & An, 2010<br />

Data was collected and analyzed by the external facilitator. Only annonymized results were made<br />

available to the researchers.<br />

3. Outcomes<br />

Outcomes of the focus group are presented in this section. Suggestions for addressing the issues are<br />

outlined in the next section.<br />

3.1 Current use of teaching and learning technologies in the university<br />

The focus group participants reported various uses of teaching and learning technologies. The most<br />

commonly used technologies are:<br />

Various Blackboard functions (blogs, discussion forums, collaborative workspaces, online testing)<br />

Video objects embedded and/or streamed online<br />

Slide presentations embedded online (for example using Camtasia)<br />

Lectopia<br />

Other technologies that are being used to varying degrees include:<br />

eBooks - to which academics can contribute comments<br />

Commercial pay per use online quizzes<br />

Pebblepad - for recognition of prior learning (ePortfolios)<br />

Wikis – both internal and external<br />

Business simulation tools<br />

Web conferencing (via Elluminate)<br />

Delicious (bookmark sharing)<br />

Google Docs (collaborative development)<br />

Facebook (for semi-formal learning)<br />

Video conferencing between campuses and also with individual students through Skype<br />

KeePad response systems<br />

iPhone Apps<br />

The focus group reported that while some academics are embedding educational technologies into their<br />

courses, there is evidence to suggest that this is not a widespread practice. Such evidence can be found<br />

in the uptake of PebblePad and Lectopia both within the business schools and the wider university.<br />

464


3.2 Industry use of information and communication technology (ICT)<br />

As universities are to create work-ready professionals, it is important to compare the exposure to, and<br />

use of technologies in the university courses to the common use of technologies in industry. There is a<br />

large range of technologies being used in industry – some of them specific to the industry itself. Of the<br />

more common technologies, the following were considered in use across a range of industries (in no<br />

particular order):<br />

Wikis<br />

Search engines<br />

Standard desktop tools: spreadsheets, document processing etc<br />

Website development tools<br />

Communications software<br />

Presentation software<br />

Skype<br />

Microblogging<br />

Social networking (eg LinkedIn, Facebook)<br />

Instant messaging<br />

Simulations<br />

Videoconferencing<br />

Content management systems<br />

Filing systems<br />

Project management tools<br />

Blogs<br />

Email<br />

Phone - SMS, Smart phones<br />

Discussion boards<br />

In comparing these to the technologies used in the university, it was considered that some academics<br />

use some of these, but many academics don’t use any of these technologies at all.<br />

3.3 Barriers to embedding teaching and learning technologies into the curriculum<br />

From the previous deliberations it was considered that there are particular barriers to academics adopting<br />

and embedding technologies on a wider scale. The group listed and considered a large range of barriers<br />

that might exist. As a small group, it was not possible to consider solutions to all of these, so the issues<br />

were sorted, grouped and a decision was made that the following are most pressing barriers:<br />

1. The value placed on teaching and issues of career progression<br />

2. Time issues<br />

3. Logistics and large class sizes<br />

4. Workplan issues<br />

Suggestions in addressing these issues are presented in the next section.<br />

Other barriers that were listed and should be considered included:<br />

Technology is used as an ‘add on’, not embedded into the course<br />

Lack of appreciation for pedagogy/instructional design<br />

Very limited understanding of available technologies and how these can support and enhance<br />

learning and teaching<br />

Kudos is given for the adoption of a technology rather than good teaching practice, neither for the<br />

effective and sustainable use of the new technology<br />

465


Many academics do not attend learning and teaching forums/PD<br />

Over focus on Good Teaching Score (GTS)<br />

<strong>Academic</strong> preferences and cultural resistance - not interested in adoption of technology<br />

Startup costs<br />

Issues of technical support for both the adoption of a new technology and on-going support.<br />

Lack of internal budget to support innovative use of technology<br />

Poor user interfaces eg previous versions of Blackboard<br />

Fear of losing IP if it is put online<br />

Access to materials is limited to one semester<br />

Universities focus on raising their international rankings impacts on what activities are supported<br />

3.4 Suggestions<br />

The group considered the four most pressing barriers to adoption and provided the following suggestions<br />

as to what the university could do to support staff in their adoption of educational technology tools in the<br />

curriculum.<br />

3.4.1 Value of teaching /career progression/motivation<br />

It is believed that teaching is not valued or rewarded by many universities, particularly in Australia where<br />

the main academic focus seems to be only on “research”. The current funding model used by the<br />

Australian Federal Government and the increased focus on raising rankings impact on what activities are<br />

supported and do not provide incentive to engage in development of teaching. Research and publication<br />

receive more recognition and facilitate progress in one’s career. This is unlike the U.S. system where one<br />

can choose a teaching career path in academia. This is a de-motivator for academics to focus on good<br />

teaching practice and continuously improve their teaching skills.<br />

Australian universities could provide support in the following ways:<br />

Teaching should be supported as a career option. This would include recognition of leadership, the<br />

scholarship of teaching, and innovation in teaching practices. Benchmarks need to be developed for<br />

good teaching practice to be measured against. The university needs to enable more flexibility by<br />

providing alternative career paths for academics, for example, teaching or research.<br />

Recognition and reward should be for developing good learning experiences not for merely adopting<br />

a new technology. The university needs to reinforce the effective use of new technologies, not just<br />

the adoption of them. Managers need to acknowledge the benefits of developing teaching skills and<br />

the value this brings to student experiences and learning outcomes.<br />

Australian universities have a strong focus on Good Teaching Score (GTS) and if innovations do not<br />

contribute to improved GTS, then they are not encouraged. As GTS is often negatively impacted in<br />

the first trial of a new technology, this is a disincentive for academics to be innovative. Universities<br />

need to redefine what constitutes good GTS and alternative recognition schemes for teaching need<br />

to be investigated. Some schools at the university have internal schemes with additional recognition.<br />

These schemes should be investigated and, if appropriate, adopted more widely throughout the<br />

university.<br />

<strong>Academic</strong>s need to develop skills in how to use technology effectively as a teaching tool.<br />

Unfortunately, many learning and teaching forums are not well attended. The university needs to find<br />

a way to reward academics for attending and contributing to learning and teaching forums and to<br />

support and encourage the development of skills in the design of technology supported learning<br />

activities.<br />

Innovation sometimes fails, however there are always lessons to be learned that can improve<br />

teaching practice and adoption of technology in the future. Universities need to develop a process for<br />

building upon these foundations and acknowledging those academics who have been early adopters.<br />

Adopting new teaching practices and technologies incurs additional time and cost. There should be<br />

discretionary budget allocated to schools to provide resources for innovations. Some schools already<br />

466


allocate small budgets for the trials of innovations. Where this model has been implemented, it should<br />

be investigated with a view to adopting it more widely within the university.<br />

Currently, academics feel that there is no recognition and support for day-to-day innovation of<br />

teaching practice. Successes are largely legitimised when there is some form of external grant<br />

associated with the development. The university needs to develop a process and culture of<br />

supporting and recognising innovation internally, that is, amongst one’s peers.<br />

There appears to be a strong cultural resistance from academics to moving away from face-to-face<br />

teaching. Technology is often viewed as an ‘extra’ which means it is viewed as taking extra time and<br />

resources. This view needs to be addressed.<br />

There appears to be a lack of understanding around copyright and IP. On the one hand, academics<br />

fear losing their IP if it is digitised, and on the other, they are not encouraged to use content which is<br />

publicly available. Universities needs to ensure that academics and copyright advisors are fully aware<br />

of their digital rights and obligations.<br />

3.4.2 Time issues<br />

Time issues relate to both academics and students and also to time pressures associated with the<br />

amount of content to be covered in various curricula. <strong>Academic</strong>s need to be supported and encouraged<br />

to be creative and experiment with technologies, that is, time allocated and no repercussions for less than<br />

successful experimentation.<br />

In this regard, universities could provide support in the following ways:<br />

It takes time to become familiar, and stay up to date, with teaching and learning technologies. It<br />

would help academics if time was allocated for this.<br />

The work plan is critical to success, therefore it should include and reward the adoption and effective<br />

use of technologies.<br />

In the early adoption phase academics need technical support. It is essential that educational<br />

designer and media production support teams understand educational needs and provide support in<br />

a timely way.<br />

As it takes time to develop resources, there should be more encouragement/incentive for academics<br />

to share resources in order to leverage better practices.<br />

To some extent, technology is being seen (and used) as an add-on to teaching and learning which<br />

merely takes more of an academic’s scarce time. Universities need to support and encourage the use<br />

of technology as an integral part of the teaching and learning process that will save time.<br />

Interestingly, and independently, these findings reflect those of McGee & Diaz (2007) who argued that<br />

under ideal circumstances, a faculty member may require anywhere from three to four terms to adopt a<br />

learning technology tool and even more time to produce positive results in teaching and learning.<br />

3.4.3 Logistics and large class sizes<br />

In addition to the time issues outlined above, many of the very large classes in business schools are<br />

compulsory, making the audience difficult to manage as they don’t really want to be there.<br />

Internationalization brings its own set of challenges and in order to service overseas students well, the<br />

use of technology is essential.<br />

Support in the form of the following can help to address these issues:<br />

Provide more staff to manage the workload. The team could include: teaching assistants,<br />

tutors/sessional lecturers and educational designers.<br />

Educational designer and media production staff who are well versed in the use of technologies<br />

should be made available to teaching staff on an ‘as needs’ basis.<br />

There is a need to understand how to build an effective teaching team for very large subjects. The<br />

university should support and encourage this development.<br />

The inflexibility of IT policy is usually a deterrent to exploring new technologies. While it is understood<br />

that security is paramount, there needs to be some leeway for academics to experiment with a range<br />

of teaching and learning technologies.<br />

467


3.4.4 Work plan issues<br />

It was considered that work plans are key to many of the issues impacting upon the adoption and<br />

embedding of technology. There are differences across schools in terms of how innovations are<br />

incorporated in the work plan.<br />

The following are suggestions for improvement:<br />

The work plan could list the range of innovations which are used and this could be used to assess<br />

academics for promotion.<br />

Work plans could be attached to the resource structure and a direct course budget.<br />

The structure of work plans should be examined with a view to making them less linear.<br />

4. In conclusion<br />

Interactive eLearning tools and web 2.0 offer great potential to engage students in richer and more<br />

interactive learning experiences that are aligned with their learning preferences. Teaching approaches<br />

can adapt their instructional strategies to take advantage of these new developments to optimize the<br />

learning outcomes for both the students and the teaching staff. A shift from content delivery and recall to<br />

facilitated interactive learning experiences is easier using new technologies, but there is a resistance<br />

amongst many teaching staff to move in this direction. Whilst some early adopters have innovated and<br />

applied some new technologies to achieve great outcome (with evidence to show this generates better<br />

capability development), these cases are still in the minority.<br />

This study looks to understand the underlying barriers to embedding the new technologies into an<br />

integrated learning experience and find pragmatic ways to remove them. It is hoped that such an<br />

approach will enhance the teaching practices and learning experiences of both teachers and students<br />

and provide graduates with stronger capabilities to perform in the workplace.<br />

Web 2.0 technologies create an environment where a teacher becomes a facilitator of learning rather<br />

than a distributor of knowledge (Williams & An, 2010). <strong>Academic</strong>s need support and encouragement to<br />

assume this new role and to use technology effectively for teaching and learning. Technology is not a<br />

generic solution and in order to implement it effectively, universities need to provide an environment that<br />

encourages and supports innovation and good teaching practice. It also needs to demonstrate<br />

commitment to the use of technology by embedding them in practice, for example, academics should be<br />

using Ed-Tools themselves to build competency and show leadership around the value of these tools.<br />

Acknowledgement<br />

The authors wish to thank Mr Stuart Whitman, managing director of ScienceAdvantage for his valuable<br />

insight through this project and Ms Carol Skyring for focus group facilitation and data analysis. The<br />

authors wish to also thank <strong>Academic</strong> Development Group—College of Business, RMIT University for<br />

financing this work through a Research in Learning and Teaching (RiLT) funding award.<br />

References<br />

Ajjan H & Hartshorne R, 2008, “Investigating faculty decisions to adopt Web 2.0 technologies: Theory and empirical<br />

tests”, Internet and Higher Education, Vol. 11, pp. 71-80. (Accessed 9 th September, 2010)<br />

Alexander B, 2006, “Web 2.0: A new wave of innovation for teaching and learning?”, EDUCAUSE Review, Vol. 41,<br />

no. 2, pp. 32-44. (Accessed 9 th September, 2010)<br />

Anderson, P. (2007), “What is Web 2.0? Ideas, technologies and implications for education”, [online], JISC<br />

Technology & Standards Watch, Retrieved from www.jisc.ac.uk/media/documents/techwatch/tsw0701b.pdf.<br />

Barnett M, Keating T, Harwook W, & Saam J, 2004, “Using emerging technologies to help bridge the gap between<br />

university theory and classroom practice: Challenges and successes”, School Sciences & Mathematics, Vol.<br />

102, no.6, pp. 299−314. (Accessed 9 th September, 2010)<br />

Berkman Center for Internet & Society at Harvard Law School, 2010, Available at: http://cyber.law.harvard.edu/<br />

(Accessed 9 th September, 2010)<br />

Brown J S & Adler R P (2008), “Minds on fire: Open education, the long tail, and learning 2.0”, EDUCAUSE Review,<br />

Vol. 43, no. 1, pp. 17-32. (Accessed 9 th September, 2010)<br />

Dalsgaard C (2006), “Social software: eLearning beyond learning management systems”, <strong>European</strong> Journal of Open,<br />

Distance and ELearning, No. 2006, available at:<br />

www.eurodl.org/materials/contrib/2006/Christian_Dalsgaard.htm (Accessed 9 th September, 2010)<br />

Daniels P (2009), “Course Management Systems and Implications for Practice”, International Journal of Emerging<br />

Technologies & Society, Vol. 7, no. 2, pp. 97-108. (Accessed 9 th September, 2010)<br />

468


Hossein Zadeh et al.<br />

Dearstyne B W (2007), “Blogs, Mashups, and Wikis: Oh my!”, Information Management Journal, Vol. 41, no.4, pp.<br />

24−33. (Accessed 9 th September, 2010)<br />

Ding, Y., Jacob, E. K., Caverlee, J., Fried, M., & Zhang, Z. (2009), “Profiling social networks: A social tagging<br />

perspective”, [online], D-Lib Magazine, 15. Retrieved from http://www.dlib.org/dlib/march09/ding/03ding.html.<br />

Hartshorne R & Ajjan H (2009), “Examining student decisions to adopt Web 2.0 technologies: theory and empirical<br />

tests”, Journal of Computing in Higher Education, Vol. 21, no. 2. (Accessed 9 th September, 2010)<br />

Hew K F, & Sim J W S (2010), “The use of weblogs in higher education settings: A review of empirical research”,<br />

Educational Research Review, Vol. 5, pp. 151-166. (Accessed 9 th September, 2010)<br />

Maloney, E. J. (2007), “What Web 2.0 can teach us about learning”, Chronicle of Higher Education, 53 (18) B26-B27.<br />

McGee P & Diaz V (2007), “Wikis and Podcasts and Blogs! Oh, My! What Is a Faculty Member Supposed to Do?”,<br />

EDUCAUSE Review, Vol. 42, no. 5, pp. 28-41. (Accessed 9 th September, 2010)<br />

Minocha S (2009), “Role of social software tools in education: a literature review”, Education + Training, Vol. 51, no.<br />

5/6, pp. 353-369. (Accessed 9 th September, 2010)<br />

Penntags project at the University of Pennsylvania (2008), Available at: http://tags.library.upenn.edu/<br />

Proserpio, L and Gioia, DA (2007), “Teaching the virtual generation”, Academy of Management Learning and<br />

Education, Vol. 6, No. 1, pp 69-80<br />

Richardson W (2009), Blogs, wikis, podcasts, and other powerful web tools for classrooms, 2nd Ed, Thousand Oaks,<br />

CA: Corwin Press.<br />

Williams K & An Y (2010), “Teaching with Web 2.0 Technologies: Benefits, Barriers and Lessons Learned”,<br />

International Journal of Instructional Technology and Distance Learning, Vol. 7, no. 3, pp. 41-48. (Accessed 9 th<br />

September, 2010).<br />

469


470


PhD<br />

Research<br />

Papers<br />

471


472


Evaluating the Performance of ERP Systems in Saudi<br />

Arabian Higher Education: A Stakeholders’ Perspective<br />

Mona Althonayan and Anastasia Papazafeiropoulou<br />

Brunel University, UK<br />

mona.althonayan@brunel.ac.uk<br />

Anastasia.Papazafeiropoulou@brunel.ac.uk<br />

Abstract: Enterprise resource planning (ERP) systems are complex and comprehensive software designed to<br />

integrate business processes and functions. Despite the difficulties and risk adopting ERP systems is expanding<br />

rapidly. Furthermore, universities make large investments in information systems (IS) expecting positive impacts<br />

(return on investment). Moreover, universities are facing serious challenges implementing a new technology. Meeting<br />

stakeholders’ expectation in higher education is one of those challenges; this refers to the universities as unique<br />

organizations. Since the effectiveness of post implementation of ERP systems has become an essential indicator of<br />

success; effective selection, development and improvement of information systems requires a systematic evaluation<br />

tool. Although there are a variety of IS success evaluation studies, there is no consensus on the appropriate manner<br />

of evaluation of IS success to help organisations return on investments in information systems. Previous studies<br />

focused on information systems and user performance, but they highlight the need for more focus on ERP systems<br />

and stakeholders’ performance. Despite the importance for IS evaluation, there is a lack of accepted frameworks on<br />

IS evaluation in general and ERPs in specific in higher education. This paper focuses on three information system<br />

models: Delone and McLean’s IS success, Task Technology Fit (TTF) by Goodhue and End User Computing<br />

Satisfaction (EUCS) by Doll and Torkzadeh. To avoid the weakness in these three models, this paper attempts to<br />

integrate the three models to produce a new construct which has a more comprehensive view of the most important<br />

factors pertinent to evaluation of stakeholders’ performance on ERP systems in higher education, combining the<br />

impact and the quality of the system. This framework will be applied at a later study the stakeholders’ performance in<br />

Saudi Arabian higher education.<br />

Keywords: ERP systems; performance; stakeholders; evaluation; higher education, individuals, framework<br />

1. Introduction<br />

Organizations spend millions of dollars on information systems to improve organizational or individual<br />

performance. According to Petter et al., (2008) information systems (IS) are developed using information<br />

technology (IT) to help individual performance. Practitioners and researchers seek to understand and<br />

measure the success of these investments (Goodhue, 1995). Despite the importance of IS success,<br />

Sedera et al., (2003) indicate the success of large IS, particularly enterprise resource planning (ERP)<br />

systems is arguably difficult to measure, since ERPs generate substantial and intangible benefits and the<br />

systems involve many users (stakeholders) ranging from top executives to data entry operators, who can<br />

define its success differently. In addition, there are many applications that span the organization, with a<br />

diversity of capabilities and functionality. Thus, measuring ERP systems’ success is a complex<br />

endeavour.<br />

ERP has played a significant role both in the recent history of information technology management in<br />

higher education, and the history of higher education itself. It is important to define the concept of ERP<br />

systems in higher education: “multiple in scope, tracking a range of activities that include human<br />

resources systems, student information systems and financial systems” (Robert et al., 2004).<br />

It is clear that there are many similarities between implementing ERP system software in the education<br />

sector and other organisations/sectors. This uniqueness is based on different combinations of certain<br />

characteristics, which were determined by Okunoye and Folick (2006) as: “complexity of purpose, limited<br />

measurability outputs, both autonomy and dependency from wider society, diffuse structure of authority,<br />

and internal fragmentation”.<br />

In higher education institutions, the stakeholders are considered a fundamental factor that distinguishes<br />

education institutes from other organizations. The major proponents of ERP systems, such as Swartz and<br />

Orgill (2000) argue that there are many encouraging reasons to implement ERP systems, for example, to<br />

improve information access and the effectiveness of workflow. Moreover, other reasons to consider ERP<br />

are capability to improve controls and the ability of stakeholders to use ERP easily. In contrast, Bradley<br />

and Lee (2007) identified that universities have similar problems to other organizations, such as<br />

coordinating resources, controlling cost and motivating and facilitating ERP amongst the faculty and the<br />

staff; therefore, IT evaluation is important, especially with the large budget universities spent on IT/ERP<br />

473


Mona Althonayan and Anastasia Papazafeiropoulou<br />

systems project. ERP systems literature has considered manufacturing industries, but few studies<br />

discuss ERP in academic institutes. Despite the rapid current growth of the ERP in higher education<br />

sector, there is lack of scholarly publication discussing ERP implementation in higher education (Rabaa`i<br />

et al., 2009).<br />

Since the value from IS evaluation and the impacts of ERP system on both the organization and<br />

individuals, is to justify the value of productivity, quality, and competitiveness of the organization, the aim<br />

of this paper is to integrate the three models D& M IS success, Task Technology Fit (TTF) Goodhue and<br />

End User Computing Satisfaction (EUCS) Doll and Torkzadeh, to produce a new construct which has a<br />

more comprehensive view of the most important factors by which to evaluate stakeholders’ performance<br />

on ERP systems in higher education.<br />

First evaluation of stakeholders’ performance is discusses on and the three selected models the D&M,<br />

TTF and EUCS are introduced. The Proposed framework including the final factors chosen from the three<br />

models to evaluate stakeholders’ performance on ERP systems in higher education is then discussed.<br />

2. Evaluation of stakeholders’ performance<br />

The following subsections discusses three IS models to evaluate ERP system stakeholders’<br />

performance.<br />

2.1 Task-Technology Fit (TTF)<br />

Task Technology Fit (TTF) is defined by Goodhue (1995) as “the extent that technology functionality<br />

matches task requirements and individual abilities”.<br />

On the other hand, Goodhue and Thompson (1995) identified TTF as “the degree to which a technology<br />

assists an individual in performing his or her portfolio of tasks”. The concept was derived originally from<br />

work adjustment theory (Dishaw and Strong, 1999).<br />

Chang (2008) explained that the TTF model considers the degree to which the capabilities of the<br />

technology match the demand of the task, and has four main constructs: task characteristics, technology<br />

characteristics, which together affect the third construct of TTF which in turn affects the outcome variable,<br />

either utilization or performance (Dishaw et al., 2002). In addition, Goodhue et al., (2000) stated that TTF<br />

presumes that the performance impacts upon the fit between three constructs: technology characteristics,<br />

task requirements, and individual abilities. The TTF model posits that IT will be used if the functions<br />

available to the user support (fit) the activities of the user.<br />

Task<br />

characteristics<br />

Technology<br />

characteristics<br />

Individual<br />

characteristics<br />

Task<br />

Technology Fit<br />

Performance<br />

Impacts<br />

Figure 1: The model of Task Theology Fit (Goodhue, 1995)<br />

Since the measurement of system success is difficult, many management information systems (MIS)<br />

researchers depend on user evaluation of systems as a surrogate for MIS success, that means the<br />

assessment made by a user (Goodhue, 1995).<br />

The TTF instrument is conceptually based on the Task Technology Fit theory in which the<br />

correspondence between information systems functionality and task requirements leads to positive user<br />

evaluations and positive performance impact (Goodhue, 1998).<br />

474


Mona Althonayan and Anastasia Papazafeiropoulou<br />

Figure 2: TTF model and user evaluation (Goodhue, 1995)<br />

Goodhue and Thompson (1995) argued that greater use leads to better performance only when there is<br />

task-technology fit. Goodhue and Thompson designed a new model to demonstrate the link between the<br />

information technology and individual performance. This connection draws on insights from two<br />

complementary researches (user attitudes as predictors of utilization and task technology fit as predictor<br />

of performance). This new model is called the Technology-to-Performance Chain (TPC). This new model<br />

emphasizes that for the information technology to have a positive impact on individual performance; the<br />

technology must be utilized and must be a good fit with the tasks it supports.<br />

Although Kositanurit et al., (2006) assert that the TTF model does not answer the question of what<br />

characteristics of the IS lead to the highest levels of user performance, it suggests some constructs that<br />

are relevant to investigation. Researchers have attempted to integrate the TTF model with other models.<br />

For instance, Dishaw and Strong (1999) conducted a study to integrate between the technology<br />

acceptance model (TAM) and TTF. They aimed to provide a theoretical basis for exploring the factors<br />

that explain software utilization and its link with user performance. Dishaw and Strong (2002) extended<br />

their work by investigating the relationship between computer self-efficacy and the combined TAM\TTF<br />

model. Continuing their effort to combine models, Dishaw and Strong (2004) integrated two models,<br />

namely, TTF and the unified theory of acceptance and use of technology (UTAUT).<br />

Gros et al., (2005) supported the combination between TTF and other models by saying it is important to<br />

consider how systems help or impede the user, but it is not sufficient alone to explain system’s success<br />

or satisfaction with it.<br />

2.2 Information Systems Success model<br />

Delone and McLean’s (1992) IS success model is widely cited, and has been a valuable contribution to<br />

the literature on IS success measurement, because it was the first study that tried to impose some order<br />

to develop a comprehensive IS model and instrument for a particular context (Gable et al., 2008).<br />

Delone and McLean (1992) conducted a large number of studies found in the academic literature<br />

covering the period of 1981-1987 attempting to identify those factors that contribute to information<br />

systems success. Based on these studies, they defined six major dimensions or categories of IS<br />

success: systems quality, information quality, use, user satisfaction, individual impact and organisational<br />

impact. The main contributions Delone and McLean make to our understanding of IS success evaluation<br />

are: (1) a method for classifying the large number of IS success measures that have been used in the<br />

literature into six categories; (2) their approach begins to identify relevant stakeholder groups in the<br />

process of evaluation; (3) they suggest a model of “temporal and causal” interdependencies between<br />

these categories (Seddon and Kiew, 1994; Myers et al., 1997; Seddon, 1997).<br />

System quality<br />

Information<br />

quality<br />

Task<br />

Technology<br />

Individual<br />

Use<br />

User<br />

Satisfaction<br />

Figure 3: D&M (1992) I /S success model<br />

User evaluations<br />

(surrogate for TTF)<br />

Task Technology Fit<br />

(TTF)<br />

475<br />

Individual<br />

impact<br />

Performance<br />

Organizational<br />

impact


Mona Althonayan and Anastasia Papazafeiropoulou<br />

Delone and McLean (1992) proposed an IS success model, but did not empirically tested it (Sabherwal et<br />

al., 2006). However, many studies have attempted to test and modify Delone and McLean’s model and<br />

called for further development and validation of their model. For instance, Seddon and Kiew (1994) were<br />

the first to empirically test the causal/process nature of the model by testing part of Delone and McLean’s<br />

model, and Seddon (1997) re-specified and extended the Delone and Mclean model and presented an<br />

alternative model of IS success. Rai et al., (2002) and Sabherwal et al., (2006) further built on Delone<br />

and McLean’s (1992) model.<br />

Seddon (1997) argued that Delone and McLean attempted to do too much by combining both the process<br />

and causal explanation of IS success in their model, and that as a result the model was confused and<br />

miss specified. The importance of Seddon’s (1997) study is that he adopted a theoretical approach to<br />

modify Delone and McLean’s model, and he distinguished between actual impact and expected impacts,<br />

and incorporated the additional construct of perceived usefulness (Sabherwal et al., 2006).<br />

Seddon (1997) believed that successful systems will provide benefits such as helping the user do more<br />

or better work in the same time, or take less time to achieve as much work of the same quality as was<br />

done in the past.<br />

The focus in this paper is on individual impact, which was defined by Delone and McLean (1992) as “the<br />

effect of information on the behaviour of the recipient of all the measure of IS success”. Gable et al.,<br />

(2008) identified individual impact as the “measure of the extent to which the (IS) has influenced the<br />

capabilities and effectiveness, on behalf of the organization, of key users”. The reason for choosing the<br />

individual impact is practicality, because impact is closely related to performance, meaning, according to<br />

Delone and McLean (1992) that impact could be an indication that the information system has given the<br />

user a better understanding of the decision context, improved his or her decision-making and productivity<br />

and has brought about a change in user activity or changed the decision-maker’s perception of the<br />

importance or usefulness of the IS. The usefulness is the degree to which a person believes that using<br />

particular systems would enhance his or her job performance (Seddon, 1997).<br />

Pitt et al., (1995) augmented Delone and McLean’s model to include service quality as a measure of IS<br />

success. They argued that Delone and McLean’s model needs to be expanded to reflect the IS<br />

department’s service role. Moreover, the basis of Delone and Mclean’s categorization theory is<br />

communication; thus, the IS department is not just a provider of products, it is also a service provider.<br />

According to Petter et al., (2008) many researchers have suggested that service quality is an important<br />

factor to be added to the Delone and McLean’s IS success model, because it is salient to IS success. In<br />

addition, there is a danger of miss -measuring IS effectiveness if researchers do not include an<br />

assessment of service quality.<br />

Petter et al. (2008) defined service quality as “the quality of the support that systems users receive from<br />

the IS department and IT support personnel”. Moreover, it measures the service quality of IT departments<br />

as opposed to individual IT applications, by measuring and comparing user expectation and their<br />

perceptions of the IT department (Petter et al., 2008). There is support for this argument in the IS<br />

literature. For example, Conrath and Mignen (1990) reported that the second most important component<br />

of user satisfaction, after general quality of service, is the match between user’s expectations and actual<br />

IS service. Rushinek and Rushinek (1986) added that fulfilled user expectations have a strong effect on<br />

overall satisfaction.<br />

Pitt et al. (1995) propose that service quality can be assessed by measuring customer expectations and<br />

perceptions of performance level for a range of service attributes. Then the difference between<br />

expectation and perceptions of actual performance can be calculated and averaged across attributes.<br />

A decade later, Delone and McLean (2003) reviewed and evaluated this argument before updating the<br />

information systems success model, based on a review of more than100 articles of the empirical<br />

conceptual literature on IS success that was published during the same period. Petter et al., (2008)<br />

explained the utility of the update to the Delone and McLean IS model and evaluated its usefulness in<br />

light of dramatic changes in IS practice, especially the explosive development of e-commerce. The<br />

update study was conducted on six dimensions: systems quality, information quality, service quality, use,<br />

user satisfaction and net benefits.<br />

476


Information<br />

quality<br />

System quality<br />

Service quality<br />

Mona Althonayan and Anastasia Papazafeiropoulou<br />

Intention to use Use<br />

User satisfaction<br />

Figure 4: Updated D&M IS success model (2003)<br />

This research focuses on service quality, which is considered an important dimension in IS measurement<br />

success, because suggests that there is a correlation between stakeholders’ expectation of the service<br />

quality and their performance level. Therefore, considering the service quality dimension for measuring<br />

stakeholders’ performance of ERP systems in higher education is really essential. Table 1; illustrates<br />

Delone and McLean model the service quality factors proposed in Pitt et al. (1995) and confirmed in,<br />

D&M update (2003).<br />

Table 1: Service quality factors, D&M information system update model (2003)<br />

Service quality<br />

-Reliability<br />

-Assurance<br />

-Empathy<br />

-Tangibles<br />

-Responsiveness<br />

Sedera and Gable (2004) attempted to build upon their previous work (2003) to derive a standardized<br />

instrument for measuring enterprise systems success, based on the research cycle developed by<br />

Mackenzie and House (1979) and McGrath (1979) that which entails two main phases: (1) exploratory<br />

phase and (2) confirmatory phase. Results from confirmatory factor analysis utilizing structural equation<br />

modelling techniques confirmed the existence of four distinct and individually important dimensions of<br />

ERP systems: individual impact, organisational impact, system quality and information quality individual<br />

impact.<br />

System<br />

quality<br />

- Ease of use<br />

- Ease of<br />

learning<br />

- User<br />

requirements<br />

- System<br />

features<br />

- System<br />

accuracy<br />

- Flexibility<br />

- Sophistication<br />

- Integration<br />

- Customization<br />

Enterprise<br />

systems<br />

success<br />

Figure 5: Sedera and Gable (2004) Validated Measures of ERP<br />

Individual<br />

Information impact<br />

quality - Learning<br />

-Availability<br />

- Usability<br />

- Understand<br />

ability<br />

- Relevance<br />

- Format<br />

- Conciseness<br />

477<br />

- Awareness/<br />

Recall<br />

- Decision<br />

effectiveness<br />

- Individual<br />

productivity<br />

Net Benefits<br />

Organization<br />

al impact<br />

- Organizational<br />

cost.<br />

- Staff<br />

requirements.<br />

- Cost reduction.<br />

- Overall<br />

productivity<br />

- Improved<br />

outcomes/outputs<br />

- Increased<br />

capacity<br />

- E-government


Mona Althonayan and Anastasia Papazafeiropoulou<br />

Petter (2008) believed that Sedera and Gable’s (2004) modification model, particularly the instrument for<br />

evaluating IS success, is unique. His point of view is based on two different reasons: (1) Sedera and<br />

Gable’s model captures the multidimensional and complex nature of IS success by measuring four<br />

dimensions; (2) another strength in this model is that the instrument was tested within the context of ERP<br />

systems to ensure its validity. In contrast, Darmawan (2001) argued that combining two or more levels<br />

into a single level analysis causes the aggregation or disaggregation of data collected at the lower<br />

(individual) to the higher (organisation) level, and could introduce bias, meaning an over- or underestimation<br />

of the significance of effects associated with variables that are aggregated or disaggregated.<br />

The reason for selecting the individual impact dimension from Delone and McLean’s IS success model<br />

(1992) and service quality impact from the update model is because both dimensions can help this<br />

research, to discover whether or not ERPs improve the stakeholders’ performance and how it meets their<br />

requirements.<br />

The reason for selecting the individual impact dimension from Delone and McLean’s IS success model<br />

(1992) and service quality impact from the update model is because both dimensions can help this<br />

research, to discover whether or not ERPs improve the stakeholders’ performance and how it meets their<br />

requirements.<br />

2.3 End user computing satisfaction<br />

The end user computing satisfaction model designed by Doll and Torkzadeh (1988) is a potentially<br />

measurable surrogate for utility in decision-making, while users interact directly with the application<br />

software to enter information or prepare output reports. The end user application’s utility in decisionmaking<br />

is enhanced when the output meets the user’s requirements (Doll and Torkzadeh, 1988).<br />

End user computing satisfaction is conceptualized as the effective attitude towards a specific computer<br />

application by someone who interacts with the application directly (Doll and Torkzadeh, 1988).<br />

The EUCS instrument consists of five factors: content, format, timeliness, ease of use/efficiency and<br />

accuracy. Although this model is focused on end user computing satisfaction, it includes factors which<br />

could be useful for IS/ERP systems measurement.<br />

EUCS was designed for a conventional computing environment. Therefore, performance was excluded,<br />

since as Doll and Torkzadeh explain, performance-related behaviours may be application specific, and<br />

development of generalizable measure of EUCS success is difficult (Amoli and Farhoomand, 1996).<br />

Content<br />

Accuracy<br />

End- User<br />

Computing<br />

Satisfaction<br />

Format<br />

Timeliness<br />

Figure 6: End user computing satisfaction model (Doll and Torkzadeh, 1988)<br />

In the ERP systems domain Somers et al., (2003), adopted the EUCS model to measure end user<br />

satisfaction in ERP systems, their study examined further the theoretical meaning, structure,<br />

dimensionality, reliability and validity of EUCS when adopted to measure EUCS with ERP software<br />

applications. Their study confirmed that the EUCS instrument can be better understood and applied as a<br />

standardized measure of advanced information technology, for instance in ERP systems application.<br />

478<br />

Ease of<br />

use


Mona Althonayan and Anastasia Papazafeiropoulou<br />

<strong>Academic</strong>s and practitioners have continually sought a reliable and valid measure of information systems<br />

success. Ideally, success measures were focused on user behaviour or measuring decision outcome,<br />

other than what users consider value in a system which is linked to how it helps them to achieve their<br />

goal (Torkzadeh et al., 2005).<br />

The selection factors which are illustrated in Table 2 are derived from the D&M model (individual impact<br />

factors), TTF model and EUCS model, and are based on their relevance to the performance and system<br />

quality. (*) Refers to the factors most related to performance, (Q) Refers to the factors most related to the<br />

quality of the system, while (R) refers to repeated factors in TTF, EUCS and D&M models.<br />

Table 2: The selection factors from the three models<br />

Information systems success factors Task technology Fit End User<br />

Computing<br />

Satisfaction<br />

-Time taken to complete a task *<br />

-Improved personal productivity *<br />

-Time efficiency of task accomplishment *<br />

-Interpretation accuracy *<br />

-Computer awareness *<br />

-Confident on performance*<br />

-User confidence*<br />

-Task performance*<br />

-Confidence in decision*<br />

-Ability to identify solutions*<br />

- Ability to identify strategic opportunities or problems*<br />

-Ability to forecast firm performance*<br />

- Time to solve problem*<br />

- Accuracy of problem solution*<br />

- Efficiency of effort *<br />

- Time savings*<br />

-Personal effectiveness*<br />

- Problem identification*<br />

-Immediate recall of information*<br />

-Delayed recall of information*<br />

3. Proposed framework<br />

-Lack of<br />

confusion(Q)<br />

-Level of detail<br />

-Locatability<br />

-Meaning<br />

-Right data(Q)<br />

-Accessibility (Q)<br />

-Assistance(Q)<br />

-Authorization(Q)<br />

-Ease of use (Q)<br />

-Flexibility (Q)<br />

-System<br />

reliability(R)<br />

-Training (Q)<br />

-Accuracy (Q)<br />

-Compatibility (Q)<br />

-Currency (Q)<br />

-Presentation<br />

-Content(Q)<br />

-Format (Q)<br />

-Timeliness(Q)<br />

-Ease of use(R)<br />

-Accuracy(R)<br />

The factors in Table 2 derived from the TTF and EUCS models, when combined together can help to<br />

evaluate the performance from the technical aspects. Moreover, the selected factors are the most<br />

suitable in the ERP systems environment, and aim to measure how the ERP systems enhance individual<br />

performance. This study excludes some of the factors in the TTF model. Goodhue (1998) asserted that<br />

presentation and level of detail have similar meaning in EUCS model, which are content and format. In<br />

contrast, locatability and meaning will not help to evaluate ERP systems from the stakeholders’<br />

performance together with the individual impact because Goodhue (1998) believed that TTF measures<br />

are intended to evaluate all systems and service of IS department, whilst the EUCS focuses on individual<br />

application.<br />

Furthermore, the derivation of the factors in Table 3‘Time taken to complete task, improve stakeholders<br />

productivity, immediate recall of information, stakeholders confidence and performance and ability to<br />

identify problem and solutions’ was initially based on a comprehensive study conducted by Delone and<br />

McLean (1992) under the dimension of individual impact. All the factors were interpreted by the<br />

researcher because they all give similar meanings. Computer awareness is an important factor to<br />

measure the stakeholders’ knowledge about the system in general and ERP systems in particular. When<br />

the individual has appropriate knowledge and experience required to use the system, better performance<br />

should result. It is important to add service quality from Delone and McLean’s updated information<br />

system success model (2003) to the factors to measure the quality of support that ERP systems<br />

stakeholders received from the IT department and its impact on the individual performance.<br />

479


Mona Althonayan and Anastasia Papazafeiropoulou<br />

Table 3: The author has gathered the factors with the same meaning into one factor<br />

Information Systems Success Factors Factors<br />

- Time taken to complete a task<br />

Time taken to complete task<br />

-Time efficiency of task accomplishment<br />

-Task performance<br />

-Improved personal productivity<br />

Improve stakeholders’ productivity<br />

- Efficiency of effort<br />

-Personal effectiveness<br />

-Effectiveness in supporting decision<br />

-Immediate recall of information<br />

Immediate recall of information<br />

-Delayed recall of information<br />

-Confident on performance<br />

Stakeholders confidence on performance<br />

-User confidence<br />

-Confidence in decision<br />

- Ability to identify strategic opportunities or problems<br />

-Ability to identify solutions<br />

- Accuracy of problem solution<br />

- Time to solve problem<br />

- Time savings<br />

Ability to identify problem and solutions<br />

-Computer awareness Computer awareness<br />

None of the three models, TTF, EUCS and D&M, discussed previously, when applied separately do not<br />

provide effective evaluation of stakeholders’ performance, since TTF and EUCS evaluate the technical<br />

aspects of the systems, and the individual impact in the D&M model focuses on the human / social<br />

aspects. However, when the three models are integrated, this will effectively evaluate the stakeholders’<br />

performance.<br />

Gable et al., (2008) argued that a holistic measure for evaluating an IS should consist of dimensions that<br />

together look both backward (impact), representing the net benefits, and forward (quality), representing<br />

the best surrogate measure of probable future impact. The combination of impact and quality represents<br />

a complete measure of the information system.<br />

Impact<br />

(impacts to<br />

date)<br />

IS impact<br />

Quality<br />

(impacts<br />

anticipated)<br />

Figure 7: Gable et al (2008) the conceptual model<br />

Therefore, to overcome the shortcomings of previous models, this research aims to integrate all the three<br />

models to create a new synthesized model which has a more comprehensive view of the most important<br />

factors that affect stakeholder performance it does so by adopting the conceptual model developed by<br />

Gable et al., (2008) which combines the impact and the quality, and selecting the appropriate factors.<br />

The outcome is the factors the study gathered from Delone and Mclean’s (1992) IS success model; it<br />

considers (impact) as a half measure. The factors gathered from the TTF model and EUCS consider<br />

(quality) as a half measure, which will be used to evaluate stakeholders’ performance.<br />

Since individual performance is an essential indicator of organizational performance, consequently,<br />

studying the impact of ERP systems on stakeholders’ performance is a significant way to assess the<br />

utility of this software on higher education and how it contributes to performance efficiency and<br />

effectiveness, as shown in Table 5.<br />

480


Mona Althonayan and Anastasia Papazafeiropoulou<br />

ERPS impact<br />

Figure 8: ERPs impact adopted from Gable et al (2008) the conceptual model<br />

Table 5: The final chosen factors from three models<br />

Information Systems Success Task-Technology Fit (TTF) End User Computing<br />

Satisfaction<br />

-Time taken to complete task<br />

-Improve stakeholders’ productivity<br />

-Immediate recall of information<br />

-Stakeholders’ confidence and performance<br />

- Ability to identify problem and solutions<br />

- Computer awareness<br />

- Service quality<br />

(Reliability Assurance-Empathy-Responsiveness-<br />

Tangible)<br />

4. Conclusion<br />

-Lack of confusion<br />

-Right data<br />

-Accessibility<br />

-Assistance<br />

-Authorization<br />

-Ease of use<br />

-Flexibility<br />

-Training<br />

-Accuracy<br />

-Compatibility<br />

-Currency<br />

-Content<br />

-Format<br />

-Timeliness<br />

Several widely used instruments have been tested for user satisfaction in IS in general and ERP systems<br />

in particular, although few studies have discussed ERP systems from stakeholders’ point of view. The<br />

literature review highlights the importance of ERP from technical and organisation levels, while social and<br />

individual levels and their aspects have been ignored. Since stakeholders are the central elements<br />

creating value through their interaction with ERP system, there is a need for studying and evaluating ERP<br />

systems by focusing on how the human factors influence success and how ERP system can improve<br />

stakeholders’ performance.<br />

The aim of this study was to develop integrated framework derived from the three widely used models:<br />

Delone and McLean’s IS success, Task Technology Fit (TTF) and End User Computing Satisfaction<br />

(EUCS), which measure different aspects of factors that have impact on individual performance on ERP<br />

system environment. This study has proposed a theoretical framework that aims to evaluate the impact of<br />

ERP systems on stakeholders’ performance.<br />

The selection of the most appropriate and related factors among these models in this study will help<br />

researchers and practitioners to evaluate the stakeholders’ performance in ERP systems by focusing on<br />

the Saudi Arabian higher education.<br />

References<br />

D&M<br />

Information systems<br />

success<br />

Task-Technology Fit (TTF)<br />

End User Computing<br />

Satisfaction (EUCS)<br />

Amoli, J. E., Farhoomand, A. (1996) “A structural model of end user computing satisfaction and user performance’’,<br />

Information &Management, 30. Pp. 65-73<br />

Bradley, J., Lee, C. (2007) “ERP Training and User Satisfaction: A case Study”, Journal of Enterprise information<br />

systems, 3 (4). Pp. 33-50<br />

481


Mona Althonayan and Anastasia Papazafeiropoulou<br />

Chang, H. H. (2008) “Intelligent agent’s technology characteristics applied to online auction task: A combined model<br />

of TTF and TAM”, Technovation, 28 Pp.564–577<br />

Darmawan, I.G.N. (2001) “Adoption and implementation of information technology in Bali’s government: A<br />

comparison between single level path analyses using PLSPATH 3.01 and Amos4 and multilevel path analyses<br />

using MPLUS2.01”, International Education Journal, 2 (4) Pp. 100-125<br />

Delone, W. H., Mclean, E. R. (1992) “Information Systems Success: The Quest for the dependent variable”,<br />

Information Systems Research, 3. (1) Pp.60-95<br />

Delone, W. H., Mclean, E. R. (2003) “The Delone and McLean model of information systems success: A ten- year<br />

update”, Journal of Management Information Systems, 19 (4) Pp. 9-30<br />

Dishaw, M.T., Strong, D. M. (1999) “Extending the technology acceptance model with task – technology fit<br />

constructs”, Information &Management, 36. Pp.9-21<br />

Dishaw, M.T. Dishaw, M.T.Strong,D., Bandy,B.D., (2002) “Extending the Task Technology Fit model with self-<br />

efficacy constructs”, Eighth American <strong>Conference</strong> of Information Systems, Pp. 1021- 1027<br />

Dishaw, M.T.Strong,D., Bandy,B.D., (2004) “The impact of task technology fit in technology acceptance and<br />

utilization model”, The Tenth American <strong>Conference</strong> on Information Systems, New York: Pp. 3306-3311<br />

Doll, W., Torkzadeh, G. (1988) “The measurement of end user computing satisfaction”, MIS Quarterly, 12. (2) Pp.<br />

250-273<br />

Gable, G., Sedera,D., Chan,T., (2008) “Re- conceptualizing Information System Success: The IS – Impact<br />

Measurement Model”, Journal of the Association for Information Systems, 9. (7) Pp. 376-408<br />

Goodhue, D. (1995) “Understanding user evaluation of information systems”, Management science, 41(12) Pp. 1827-<br />

1995<br />

Goodhue, D. Thompson, R. (1995) “Task – technology Fit and individual Performance”, MIS Quarterly, 19(2) Pp.<br />

213-236<br />

Goodhue, D. (1998) “Development and Measurement Validity of a Task Technology Fit Instrument for User<br />

Evaluations of Information Systems”, Decision Sciences, 29 (1) Pp. 105- 138<br />

Goodhue, D., Klein,B.D., March,S.T., (2000) “User evaluation of IS surrogates for objective performance”,<br />

Information and Management, 38. Pp.87-101<br />

Gros, C.D., Mueller,H., Christian,L., (2005) “Evaluating user interaction with clinical information systems: A model<br />

based on human computer interaction models”, Journal of Biomedical informatics, 38. Pp.244-255<br />

King, P., (2002) “The promise and performance of enterprise systems for higher education”, Research Study from the<br />

EDUCUSE Center for Applied Research.: Pp. 1-7<br />

Kositanurit, B. Ngwenyama,O., Osei-Bryson,KM.,(2006) “An exploration of factors that impact individual performance<br />

in an ERP environment: an analysis using multiple analytical techniques”, <strong>European</strong> Journal of Information<br />

Systems, 15 .Pp. 556-568<br />

Myers, B. (1997) “A comprehensive model for assessing the quality and productivity of the information systems<br />

function: Toward a theory for information systems assessment”, Information Resources Management Journal,<br />

10. 1. Pp.6-25<br />

Okunoye, A., Folick, M. (2006) “ERP Implementation in higher education: An Account of Pre- Implementation and<br />

implementation Phases”, Journal of Cases on Information Technology, 8 (2) Pp.110- 132<br />

Petter, S., Delone,W., McLean,E., (2008) “Measuring information systems success: model, dimension, measures,<br />

and interrelationships”, <strong>European</strong> Journal of Information Systems, 17.Pp.236-263<br />

Pitt, L F., Watson, R., Kavan, C.B., (1995) “Service Quality: A measure of information systems effectiveness”, MIS<br />

Quarterly, 19.2. Pp.173-187<br />

Rabaa`i, A., Bandara, W., Gable, G., (2009) “ERP system in the higher education sector: A descriptive case study”,<br />

20 TH Australian <strong>Conference</strong> on Information Systems: Melbourne, Pp. 456-470<br />

Sabherwal, R., Jeyaraj, A., Chowa, C., (2006) “Information Systems: individual and organizational determinants”,<br />

Management science, 52. (12) Pp. 1849-1882<br />

Seddon, P. (1997) “A respecification and extension of the Delone and Mclean of IS success”, Information Systems<br />

Research, 8 (3) Pp. 240-253<br />

Seddon, P., Staples,S., Patnayakuni,R., Bowtell,M., (1999) “Dimension of Information Systems success”,<br />

Communication of the Association for Information Systems, 2 (20) Pp. 2-60<br />

Sedera, D., Gable, G., Cgan, T., (2003) “Measuring Enterprise System Success: A Preliminary Model” America<br />

<strong>Conference</strong> on Information Systems (AMCIS) Pp. 476-485<br />

Sedera, D. Gable, G. (2004) “A factor and structural equation analysis of the enterprise system success<br />

measurement model”, Twenty- fifth International <strong>Conference</strong> on Information Systems, Pp. 449-464<br />

Somers, T.M., Nelson, K., Karimi, J.,(2003) “Confirmatory Factors Analysis of the End User Computing Satisfaction<br />

Instrument: Replication within an ERP Domain”, Decision Sciences, 34 (3) Pp. 595-621<br />

Swartz, D., Orgill, K. (2000) “Higher education ERP: Lessons Learned”, EDUCAUSE 2000 in Nashville: Pp. 1-12<br />

Torkzadeh, G., Xenophon, K., Doll, W., (2005) “Confirmatory factor analysis and factorial invariance of the impact of<br />

information technology instrument”, OMEGA, 33. Pp. 107-118<br />

Torkzadeh, G., Doll, W.J. (1999) “The development of a tool for measuring the perceived Impact of information<br />

technology on work”, OMEGA, 27. Pp.327-339<br />

482


The Benefits of ICT Adoption: An Empirical Study of Nigerian<br />

SMEs<br />

Idisemi Apulu and Ann Latham<br />

University of Wolverhampton, UK<br />

i.apulu@wlv.ac.uk<br />

A.Latham@wlv.ac.uk<br />

Abstract: The successful operation of companies in most industries is becoming increasingly dependent on their<br />

ability to adopt and utilize ICT systems (Chibelushi, 2008). The rapid development of ICT is suddenly changing the<br />

conventional way of doing business in organizations. In recent times, organizations of all types are adopting ICT in<br />

order to improve organizational efficiency, productivity, communication and to strengthen their competitiveness. In<br />

the present knowledge economy, there is a need for organizations, including SMEs, to develop some forms of<br />

competitive advantage with the use of ICT in order to be successful. SMEs play a vital role in the Nigerian economy<br />

and greatly contribute to the country’s industrial development. Despite the growing number of studies on the benefits<br />

of ICT adoption in SMEs, the benefits of ICT adoption in the Nigerian context is still under researched. Thus, this<br />

paper is drawn from on-going PhD research and aims at investigating the benefits that are derived from the adoption<br />

of ICT within Nigerian SMEs. The analyses of data collected for this study would provide valuable information to SME<br />

owner-managers, policy makers and academic researchers.<br />

Keywords: ICT, SMEs, adoption, benefits, Nigeria<br />

1. Introduction<br />

In recent years, the growth of ICT has had a substantial impact on the way organizations function. Sigala<br />

(2003), states that the widespread use of ICT is changing the way people and companies work. ICT can<br />

be viewed as a collective term for a wide range of software, hardware, telecommunications and<br />

information management techniques, applications and devices, and are used to create, produce, analyse,<br />

process, package, distribute, receive, retrieve, store and transform information (Barba-Sánchez et al.,<br />

2007). Also, Gupta et al. (2008) define ICT as technologies such as the Internet, Intranets, Extranets,<br />

ERP and other such technologies that cover a wide spectrum from basic infrastructure implementation to<br />

technologies that improve services and operations in organizations. ICT helps to support human activities<br />

that enhance organizational or personal efficiency and effectiveness (Cohen et al., 2002). ICT helps to<br />

execute activities faster, support autonomous decision making processes, and enable distributive<br />

operations in order to become more efficient (Pokharel, 2005). It has revolutionized the manner in which<br />

firms interact and do business. ICT has a valuable potential for developing SMEs as the effective use and<br />

better integration of ICT in business processes could assist them to make more efficient decisions<br />

relevant to their performance (Sigala, 2003; Pokharel, 2005; Ongori and Migiro, 2010). Although there is<br />

a growing number of studies on the benefits of ICT adoption in SMEs, the benefits of ICT adoption in the<br />

Nigerian context as the area of study is still under researched. The paper is drawn from on-going PhD<br />

research and aims at investigating the benefits that are derived from the adoption of ICT within Nigerian<br />

SMEs. The analyses of data collected for this study would provide valuable information to SMEs ownermanagers,<br />

policy makers and academic researchers.<br />

2. Information and Communication Technology (ICT)<br />

Fulantelli and Allegra (2003), state that ICT offers enterprises a wide range of possibilities for improving<br />

their competitiveness and provides mechanisms for getting access to new market opportunities and<br />

specialized information services in organizations. According to Apulu and Latham (2009a), ICT helps<br />

organizations to exchange information and build closer relationships with their customers, suppliers and<br />

business partners. It also helps companies to provide immediate customer feedback that allows them to<br />

react fast to customer demands and recognize new market niches. Fulantelli and Allegra (2003) argue<br />

that the use of ICT brings about transparency in business processes in terms of dealing with<br />

stakeholders, which in turn, could lead to the development of better business practices that will help to<br />

meet customers’ service levels. ICT is an effective tool that can be used for improving external<br />

communications and delivering quality service to customers. It improves a firm’s ability to plan better for<br />

the future by providing the management with easy access to relevant information about different activities<br />

and/or operations within and outside the firm (Parida et al., 2009). Akkeren and Cavaye (1999) also state<br />

that ICT improves the ability for SMEs to compete with larger organizations and enables them to operate<br />

on an international scale. In the present knowledge economy, there is a need for organizations including<br />

SMEs, to develop some forms of competitive advantage with the use of ICT.<br />

483


Idisemi Apulu and Ann Latham<br />

3. Small and Medium Sized Enterprises (SMEs)<br />

SMEs form a significant proportion of the economy in many countries (Sarosa and Underwood, 2005).<br />

Wielicki and Arendt (2010) state that SMEs’ importance is magnified by the fact that for any country,<br />

regardless of its stage of economic development, SMEs generate the bulk of economic output (usually<br />

well over 90%). SMEs play a vital role in the Nigerian economy and greatly contribute to the country’s<br />

industrial development as they assist in creating jobs thereby improving the socio-economic development<br />

of the country. Nigerian SMEs are known as cornerstones responsible for Nigeria’s economic growth and<br />

stability (Ojukwu, 2006). They are also viewed as sources of flexibility and innovation that makes<br />

significant contributions to economies, both in terms of the number of SMEs and the proportion of the<br />

labour force employed by these firms. Lawrence (2008), states that SMEs contribution to the economy of<br />

many countries cannot be over emphasized.<br />

The definition of SMEs varies from country to country but is often based on employment, assets or a<br />

combination of the two. The Small and Medium Sized Development Agency of Nigeria (SMEDAN)<br />

defines SMEs based on the following criteria: a micro enterprise as a business with less than 10 people<br />

with an annual turnover of less than 5 million Naira, a small enterprise as a business with 10-49 people<br />

with an annual turnover of 5-49 million Naira and a medium enterprise as a business with 50-199 people<br />

with an annual turnover of 50-499 million Naira. In Nigeria, SMEs cover the entire range of economic<br />

activity within all sectors (SMEDAN, 2005). Zhouying et al. (2009), state that the use of ICT in SMEs can<br />

offer SMEs several benefits that could assist in increasing their capacities and competitiveness.<br />

4. Benefits of ICT in SMEs<br />

ICT has a significant positive impact on organizational performance (Maldeni and Jayasena, 2009) and is<br />

vital to SMEs. Without the use of ICT, modern businesses are not possible as ICT has a significant<br />

impact on SMEs’ operations and is claimed to be crucial for the survival and growth of economies in<br />

general (Berisha-Namanil, 2009). ICT provides opportunities for business transformations (Chibelushi,<br />

2008) and provides SMEs the opportunity to conduct business anywhere (Jennex et al, 2004). Brady et<br />

al. (2002), state that ICT plays an important role in enhancing the productivity and effectiveness of certain<br />

activities or functions made by SMEs. ICT can create new, strong linkages between internal activities,<br />

and even coordinate these actions more closely with their consumers and suppliers to facilitate<br />

integration within the company (Leenders and Wierenga, 2002). According to Prasad et al. (2009), ICT<br />

enhances companies’ abilities to coordinate activities regionally, nationally and globally and creates<br />

interrelationships between companies. It also expands the scope of industries in which companies<br />

compete to achieve competitive advantage (Porter and Millar, 1985).<br />

Love et al. (2004) ascertain that the use of ICT offers many benefits to SMEs at different levels<br />

(operational level, tactical level and strategic level). The benefit of ICT cuts across all sectors of the<br />

economy and all the fields of human activities. The <strong>European</strong> Commission (2008), states that SMEs<br />

could use ICT to grow and to become more innovative. Swift (2009) argues that SMEs benefit from the<br />

use of ICT as it connects them more easily and cheaply to external contacts. Through the use of ICT,<br />

SMEs can engage in e-commerce which will aid them in increasing their efficiency in their day-to-day<br />

business operations and sustain their business growth through the opening of new market channels<br />

(Ongori and Migiro, 2010). ICT enables SMEs to have access to robust business information that leads<br />

to organizational effectiveness (Irani, 2002). Akkeren and Cavaye (1999) state that ICT provides a cost<br />

effective way for SMEs to market their business, launch new products, improve communications, gather<br />

information and identify potential business partners. Hence, ICT can be regarded as a vital resource that<br />

enhances the competitiveness of SMEs in any business environment.<br />

5. Methodology<br />

The research question for this study is “Does ICT help SMEs in Nigeria?” The study adopts a case study<br />

approach and aims at determining whether or not ICT helps SMEs in Nigeria. Case study research<br />

enables the researcher to discover more effectively the context of a situation. Case studies assist to gain<br />

a deeper understanding of the studied phenomenon (Yin, 2003). In this study, case studies utilizing semistructured<br />

interviews, observation, and document review were employed. Multiple sources of data were<br />

used to address the ethical need to increase the reliability and validity of the research processes (Yin,<br />

2003). According to Orlikowski and Baroudi (1991); Remenyi and Williams (1996), case study is one of<br />

the most frequently used research methods in information systems research. Also, Sauer (1993) is of the<br />

opinion that research in information systems is best done by case study.<br />

484


Idisemi Apulu and Ann Latham<br />

Semi structured interviews that focused on the in-depth understanding of the benefits of ICT adoption<br />

within SMEs in Nigeria were used as tools for collecting data during the study and enabled the<br />

researchers to elicit respondents views and experiences in their own words. 25 SMEs in Apapa, Lagos<br />

were selected in accordance with the definition of SMEs in Nigeria (10 to 199 employees) and based on<br />

their decision to deploy ICT in their business processes. All the 25 interviewees were either ownermanagers,<br />

heads of departments or IT officers. Interviews were carried out using an interview guide, with<br />

each interview lasting between 1-1½ hours and tape recorded.<br />

Although the interviews were used in order to aid better interpretation of results, the use of documents<br />

and observations were also important. The researcher examined relevant documents (e.g. annual<br />

reports, project reports) that were gathered from the participating organizations. In addition, observing the<br />

management operations and procedures of the SMEs assisted in providing better information than just<br />

relying upon company reports and key informants. This multi-method approach enabled triangulation to<br />

take place. The data analysis process of the case study involved identifying patterns in the case study<br />

data. This includes arranging data in a chronological order, writing up the data according to phases and<br />

themes, in addition to grouping the paragraphs of the same themes and phases together.<br />

Furthermore, Lagos was chosen for the purpose of this study as it is considered the commercial nerve<br />

centre in Nigeria because of its strategic location, peculiar demographics and contribution to national<br />

Gross Domestic Product (GDP). According to Lawal (2002), over 60% of industries are located in Lagos<br />

hence; Lagos could serve as the industrial characteristics of Nigeria.<br />

The primary details of the SMEs that participated in the interviews are shown in the table below.<br />

Table 1: Primary details of the SMEs that participated in the interviews<br />

SME (Case study) Type of Business Number of<br />

Employees<br />

Business Age<br />

1 Engineering 43 (Small firm) 12 years<br />

2 Medical/Health 20 (Small) 13 years<br />

3 Telecommunication/Communication 26 (Small) 18 years<br />

4 Wholesale/Retail 90 (Medium) 9 years<br />

5 Hospitality 120 (Medium) 5 years<br />

6 Telecommunication/Communication 16 (Small) 3 years<br />

7 Telecommunications/Communication 16 (Small) 3 years<br />

8 Transport/Clearing and Forwarding/Haulage 15 (Small) 2 years<br />

9 Financial Service/Service Industry 35 (Small) 5 years<br />

10 Manufacturing 80 (Medium) 10 years<br />

11 Transport/Clearing and Forwarding/Haulage 15 (Small) 21 years<br />

12 Transport/Clearing and Forwarding/Haulage 15 (small) 9 years<br />

13 Others 25 (Small) 2 years<br />

14 Communication/Telecommunication 65 (Medium) 8 years<br />

15 Manufacturing 89 (Medium) 7 years<br />

16 Financial service or service industry 25 (small) 4 years<br />

17 Pharmacy/Medical/Health 13 (small) 14 years<br />

18 Others 18 (small) 8 years<br />

19 Oil and Gas/Maritime 14 (small) 2 years<br />

20 Oil and Gas/Maritime 30 (small) 5 years<br />

21 Engineering 29 (Small) 11 years<br />

22 Financial Service/Service Industry 22 (small) 17 years<br />

23 Education 29 (small) 7 years<br />

24 Legal Practice 17 (small) 5 years<br />

25 Oil and Gas/ Maritime 15 (small) 1 year<br />

Interviewees were asked to comment on the benefits derived from the use of ICT solutions in their<br />

various firms. Respondents identified several benefits of using ICT which include:<br />

5.1 Efficiency<br />

Thirteen respondents identified efficiency as a benefit gained from adopting ICT. The respondents stated<br />

that ICT brings about efficiency in terms of conducting their various businesses. For example one<br />

respondent states that:<br />

485


Idisemi Apulu and Ann Latham<br />

“ICT has helped us to become more efficient in terms of carrying out our duties as a small<br />

company. In the past, we were using bill cards to get our stock update, which was a manual<br />

process and usually takes some days. However, since we started using ICT we can get our<br />

stock update under some hours and it has made our company to become more efficient”.<br />

5.2 Speed<br />

Some interviewees were also of the opinion that ICT brings about speed. According to one respondent:<br />

“ICT now makes us to deliver quality service to our customers within a very short time.<br />

Unlike the manual method, the use of ICT makes our work to move very fast”.<br />

5.3 Productivity<br />

Increase in productivity was identified as another benefit of using ICT in the companies. Respondents<br />

believe that the adoption of ICT in their various companies has enabled them to become more<br />

productive. According to one company:<br />

“With ICT our business is more productive in terms of meeting our targets and ICT has<br />

helped us to become more creative”.<br />

5.4 Communication<br />

Some respondents stated that ICT improves communication. According to one respondent:<br />

“ICT helps us to communicate, in terms of exchanging information. ICT is mainly used in our<br />

organization for communication with our customers, suppliers, to get quotation for various<br />

products by searching the internet and also to get information from other stake holders in the<br />

business”.<br />

Another respondent states that:<br />

“ICT helps us to communicate better with our clients. We usually receive feedbacks from our<br />

clients within a short time via email and it also helps us to manage our data”.<br />

5.5 Easy Access to Information<br />

Easy access to information was described by respondents as a benefit of using ICT. Respondents<br />

believe that the use of ICT has enabled information to become readily available in their various<br />

companies. One respondent commented:<br />

“ICT makes it very easy for us to access information. ICT is a key to success in every<br />

organization”.<br />

Another respondent states that:<br />

“ICT brings about the availability of data, that is, information availability on a daily basis”.<br />

5.6 Awareness and Increase in profit<br />

Respondents identified some other benefits of using ICT such as helping to create awareness and<br />

assisting their companies to make or increase profit. One company comments:<br />

“ICT is interesting, it has added a lot to the body of knowledge. ICT helps organizations to<br />

create awareness because it provides the opportunity for companies to advertise their<br />

products and services online. ICT has helped us to save cost and also, to increase profit.<br />

Virtually every aspect of the whole life is now embracing ICT which makes it more lucrative<br />

for companies to get more benefits and gain, such as increase in their annual returns. ICT is<br />

wonderful”.<br />

Another company comments:<br />

“ICT has helped to increase our company’s profit. Every organization is a profit oriented<br />

organization and it implies that we are here to make profit through the use of modern<br />

technologies”.<br />

486


5.7 Customer satisfaction<br />

Idisemi Apulu and Ann Latham<br />

All the respondents’ mentioned that ICT helps to provide better customer service to their customers<br />

thereby increasing customers’ satisfaction. One respondent commented:<br />

“ICT helps us to satisfy our customers by providing them with their required service”.<br />

Also, another respondent commented:<br />

“ICT is what is in vogue nowadays. You cannot do without technology. If you are competing<br />

with other organizations that are ICT compliant, it is very important that you are also able to<br />

compete, as this will help to satisfy your customers. Information has a lot to do with<br />

organizational advantage in terms of customers’ satisfaction”.<br />

5.8 Competitive advantage<br />

Eighteen respondents believe that one of the benefits of using ICT is the fact that it helps organizations to<br />

have some form of competitive advantage. According to one respondent:<br />

“The adoption of ICT in this company has enabled us to have some sort of competitive<br />

advantage against our competitors”.<br />

The other respondents also made similar comments while identifying the benefits of using ICT in their<br />

different companies.<br />

5.9 Customer growth and good reputation<br />

Two respondents stated that ICT has assisted them to attract more customers and to establish good<br />

reputation with their existing customers. One of the respondents commented:<br />

“The world globally is changing; technology is now at the forefront of every organisation.<br />

There is no way as an SME; we can meet up with the current trend of technology if we do<br />

not use ICT because everything is going digital. As a company, using manual systems<br />

nowadays can hinder good patronage from people. Hence, we had to come out and do what<br />

others are doing in order to meet up with the standard of others. The use of ICT has assisted<br />

us to attract more customers and has also helped to build a good reputation for this<br />

company.<br />

5.10 Planning and focusing<br />

One company stated that ICT helps them to plan and be more focused. ICT also helps to provide their<br />

business with the services they require. The company, a telecommunication company, comments:<br />

“The benefits of using ICT are much and based on my opinion, ICT is indispensible. For<br />

example if you do not have a computer, you cannot produce things like recharge cards.<br />

Companies are also doing away with manual accounting these days. They are moving from<br />

paper ledger to automated ledger. It is only with the use of ICT infrastructures that we can<br />

reasonably handle these developments and put things in a good format. ICT helps us to plan<br />

and to be more focused. Also, ICT helps us in the area of database management and<br />

inventory production. It constitutes the life wire of the efficient management of our business”.<br />

5.11 User friendliness<br />

User friendliness was also identified by a respondent as a benefit of ICT. The respondent stated that:<br />

“ICT is user friendly. For example as a construction company, we can design an estate<br />

within a week using ICT applications such as AutoCAD”.<br />

5.12 Confidence<br />

Another respondent believes that ICT brings confidence to business.<br />

According to the respondent:<br />

“One of the benefits we have derived from the use of ICT applications is the fact that it helps<br />

to build confidence in our customers especially when they see the use of modern<br />

technologies in this company. People can also do business with us online and this increases<br />

487


Idisemi Apulu and Ann Latham<br />

confidence. When a company have a website for example, it brings about confidence.<br />

People can go to the website and get the information they require. ”<br />

The table below provides a summary of the Findings.<br />

Table 2: Summary of the findings<br />

Benefits of ICT in Nigerian SMEs Number of Respondents<br />

Efficiency 13<br />

Speed 17<br />

Productivity 3<br />

Communication 3<br />

Easy access to information 7<br />

Awareness and Increase in profit 4<br />

Customer satisfaction 25<br />

Competitive advantage 18<br />

Customer growth and reputation 2<br />

Planning and focusing 1<br />

User friendliness 1<br />

Confidence 1<br />

6. Discussion<br />

The evidence demonstrates that there are several benefits associated with the use of ICT in SMEs. The<br />

responses obtained from the interviewees have enabled the researchers to identify some major benefits<br />

associated with the use of ICT in Nigerian SMEs. The introduction of ICT in SMEs can bring about<br />

modifications in the way they work. On the whole, ICT applications can provide several benefits across a<br />

wide range of intra- and inter- firm business operations and transactions. Parida et al. (2009), in their<br />

research with Swedish firms, discovered that ICT helps firms to access information. The authors stated<br />

that easy access to information is precious for firms operating in a dynamic environment and that the<br />

firms need to be updated on new technology, innovations amongst others, which might influence a firm’s<br />

future competitiveness. According to Ion and Andreea (2008), with ICT high efficiency is obtained in<br />

business transactions due to a fast and accurate processing of information. ICT applications can assist to<br />

improve information and knowledge management within an organization, can reduce transaction costs<br />

and can increase the speed and reliability of transactions in any organization (Apulu and Latham, 2009b).<br />

More specifically, Barba-Sánchez et al. (2007), argue that SMEs can obtain a wide range of benefits<br />

from the use of ICT. They also state that ICT can reduce business costs, improve productivity and<br />

strengthen growth possibilities.<br />

Based on the case studies and literature review, it can be said that ICT helps to improve customer<br />

service in SMEs. Parida et al. (2009), state that ICT is a tool that provides customers with better services<br />

in terms of quicker responses and closer interactions. ICT also improves communication and maintain<br />

close relationships with business partners. Most of the respondents in the study stated that ICT brings<br />

about competitive advantage. Apulu and Latham (2009b) advocate that appropriate use of ICT can assist<br />

SMEs gain competitive advantage by reducing costs and improving core business processes. Therefore,<br />

from the literature review and case studies, it is certain that one of the major benefits derived from the<br />

use of ICT in Nigerian SMEs, is the fact that it helps them to remain competitive.<br />

Some respondent also mentioned that ICT helps to increase efficiency in their various organizations.<br />

According to Cohen et al. (2002), ICT plays a supportive role in terms of enhancing organizational<br />

efficiency and effectiveness. Fink and Disterer (2006) also argue that ICT helps organizations to become<br />

more efficient, effective and competitive. Furthermore, Pavic et al. (2007) state that organizations have<br />

the opportunity to achieve a competitive advantage from the advances in ICT through market efficiency<br />

gains, better quality, innovation and customer responsiveness.<br />

Increase in profit is identified as another benefit that is derived from adopting ICT in Nigerian SMEs.<br />

Respondents stated that ICT is used as a marketing tool to help increase profit. Also some respondents<br />

mentioned that ICT is a modern technology and it is impossible for companies to cope without utilizing<br />

some form of modern technologies especially in terms of competing with other organizations that are ICT<br />

compliant. According to Ongori and Migiro (2010), the impact of globalization has enabled companies to<br />

build strong relationships with their counterparts from any part of the world by using ICT. This has obliged<br />

many SMEs to adopt ICT in order to survive in the present competitive era.<br />

488


Idisemi Apulu and Ann Latham<br />

Easy access to information was described as another benefit. Respondents stated that the use of ICT<br />

has enabled information to become readily available in their various companies. Similarly, Irani (2002)<br />

ascertains that ICT helps SMEs to have access to robust business information that leads to<br />

organizational effectiveness. Also, Funteli and Allegra (2003) argue that organizations can exchange<br />

real-time information and build closer relationships with customers, suppliers and business partners with<br />

the use of ICT. Customers can also receive immediate feedback from companies and move from manual<br />

data processing to automated data processing. Improved communication was also described by<br />

respondents as a benefit of using ICT within Nigerian SMEs. It was noted that most SMEs decided to<br />

adopt ICT to enable them communicate with their customers, suppliers, and stakeholders in their field of<br />

business. According to Kollberg and Dreger (2006), ICT provides new ways of storing, processing,<br />

distributing and exchanging information within companies and customers. Also, Akkeren and Cavaye<br />

(1999) state that ICT helps to improve communication, gather information and identify potential business<br />

partners in organizations.<br />

Ashrafi and Mutarza (2008), argue that organizations of all type around the globe are currently utilizing<br />

ICT in order to provide better customer service. In terms of customer satisfaction, all the respondents in<br />

the study stated that ICT has enabled them to provide better customer service to their customers.<br />

Respondents also mentioned that ICT has assisted them to fully compete with their competitors.<br />

According to Chowdbury and Wolf (2003), ICT helps to increase business competitiveness and enhance<br />

enterprise performance through indirect cost savings such as labour cost and increased labour<br />

productivity and direct cost such as reduction of firms input. From the case studies, it was identified that<br />

the use of ICT increases productivity as it helps SMEs to meet their desired targets and also increases<br />

the speed of delivery in terms of responding to customers’ demands. All these help to bring about<br />

customers’ satisfaction. Jones et al. (2003), state that ICT brings quick response time to markets,<br />

customers, suppliers, increases flexibility and reduces delivery time and processing of payment.<br />

Having online presence has been identified by some respondents as another benefit of adopting or using<br />

ICT in their companies as ICT provides an opportunity for businesses to offer products and services in<br />

the global markets. Some respondents also commented that having online presence helps to create<br />

awareness. For example, a respondent stated that one of the benefits of using ICT is the fact that it helps<br />

to build their customers confidence in company, especially with the use of modern technologies.<br />

Therefore, this implies that ICT helps to inform potential customers’ about a company’s service.<br />

ICT is being applied in a wide range of areas within many organisations in recent times. This has radically<br />

changed the manner in which businesses are conducted in organizations (Cohen and Kallirroi, 2006).<br />

According to Zhouying et al. (2009), ICT can help firms to overcome spatial and institutional barriers, cut<br />

time, cost of production, increase flexibility, take new marketing channel and employ new ways of doing<br />

business. This means that ICT brings about increased competitiveness.<br />

Furthermore, Parida et al. (2009) argue that ICT improves a firm’s ability to plan better for the future. A<br />

company stated that ICT helps them not only to plan and become more focused, but also to provide their<br />

business with the services they require. Two other companies mentioned that ICT helps them to build a<br />

brand and to satisfy their customers thereby assisting them to build a good reputation. In order words, the<br />

integration of ICT in SMEs business processes can greatly improve their performance.<br />

7. Conclusion<br />

The findings of this study imply that ICT helps SMEs in Nigeria. Some benefits associated with the use of<br />

ICT in Nigerian SMEs have been identified, which support the findings of other research. For example,<br />

ICT enhances organizational speed/efficiency (Cohen et al., 2002; Sigala, 2003; Pokharel, 2005; Ongori<br />

and Migiro, 2010), improves the ability for SMEs to compete (Akkeren and Cavaye, 1999; Fulantelli and<br />

Allegra, 2003), enhances productivity (Brady et al., 2002), enables SMEs to have easy access to robust<br />

information (Akkeren and Cavaye, 1999; Irani, 2002), helps to plan for the future (Parida et al., 2009) and<br />

improves communication (Akkeren and Cavaye, 1999). Furthermore, the study identifies some distinctive<br />

benefits that are associated with the use of ICT in Nigerian SMEs, such as customer satisfaction,<br />

customer growth and reputation and confidence in business.<br />

The identified benefits show that the deployment of ICT applications in these SMEs can assist in<br />

providing better customer service and enable them to effectively compete with their competitors. Based<br />

on the literature review and the case studies, it is certain that the utilization of ICT in Nigerian SMEs<br />

would assist in increasing their competitiveness. Therefore, it can be concluded that Nigerian SMEs<br />

489


Idisemi Apulu and Ann Latham<br />

stand to gain some added advantage from the use of ICT as it can help to further develop their business<br />

processes. The findings of this study could encourage SME owner-managers and IT professionals to<br />

intensify their efforts in deploying ICT in their various organizations as the study has provided some<br />

explanation on the benefits associated with the deployment of ICT within SMEs using the Nigerian<br />

context.<br />

References<br />

Apulu, I. and Latham, A. (2009a) “Knowledge Management: Facilitator for SMEs competitiveness in Nigeria”, UKAIS<br />

2009 <strong>Conference</strong>, St Anne’s College, Oxford United Kingdom.<br />

Apulu, I. and Latham, A. (2009b) “A review of factors affecting ICT adoption in SMEs in Developing Countries: a<br />

Case Study of Nigeria”, Proceedings of the ISBE 2009 <strong>Conference</strong>, University of Liverpool, United Kingdom.<br />

Ashrafi, R. and Murtaza, M. (2008) “Use and Impact of ICT on SMEs in Oman”, Electronic Journal Information<br />

Systems Evaluation, Vol 11, No. 3, pp 125-138. [online] http://www.ejise.com/volume-11/volume11issue3/ashrafiAndMurtaza.pdf<br />

[Accessed on 10th October, 2009]<br />

Barba-Sánchez, V., Martínez-Ruiz, M and Jiménez-Zarco, A. (2007) “Drivers, Benefits and Challenges of ICT<br />

Adoption by Small and Medium Sized Enterprises (SMEs): A Literature Review, Problems and Perspectives in<br />

Management, Vol 5, No. 1, pp 103-114.<br />

Berisha-Namanil, M. (2009) “The role of information technology in small and medium sized enterprises in Kosova,<br />

Fulbright Academy 2009 <strong>Conference</strong> Small Places Can Change the World. [online] Availble:<br />

http://www.fulbrightacademy.org/file_depot/0-<br />

10000000/2000030000/21647/folder/82430/Berisha+Paper+IT+in+SMEs+in+Kosovo.pdf [July 15, 2009].<br />

Brady, M.; Saren, M. and Tzokas, N. (2002) “Integrating Information Technology into Marketing Practice – The IT<br />

Realize of Contemporary Marketing Practice”, Journal of Marketing Management, Vol 18, pp 555-577.<br />

Chibelushi, C. (2008) “Learning the hard way? Issues in the adoption of new technology in small technology oriented<br />

firms”, Education + Training, Vol 50, Issue 8/9, pp 725-736.<br />

Chowdhury, S. K., and Wolf, S. (2003) “Use of ICTs and the economic performance of SMEs in East Africa”,<br />

Discussion Paper No. 2003/06.<br />

Cohen, S. and Kallirroi, G. (2006) “e-Commerce Investments from an SME perspective: Costs, Benefits and<br />

Processes”, The Electronic Journal Information Systems Evaluation, Vol 9, Issue 2, pp 45 – 56.<br />

<strong>European</strong> Commission (2008) “Making SMEs more competitive” [online] Available:<br />

http://ec.europa.eu/enterprise/sme/competitive_en.htm [October 30, 2008].<br />

Fink, D., and Disterer, G. (2006) “International case studies: to what extent is ICT infused into the operations of<br />

SMEs”, Journal of Enterprise Information, Vol. 19, 608-624.<br />

Fulantelli, G., and Allegra, M. (2003) “Small company attitude towards ICT based solutions: Some key-elements to<br />

improve it”, Educational Technology and Society, Vol 6, Issue 1. [online] Available<br />

http://www.ifets.info/journals/6_1/fulantelli.html [March 10, 2009].<br />

Gupta, B., Dasgupta, S., and Gupta, A. (2008) “Adoption of ICT in a government organization in a developing<br />

country: an empirical study”, The Journal of Strategic Information Systems, Vol 17, No 2, pp140-154.<br />

Ion, P., and Andreea, Z. (2008) Use of ICT in SMEs management within the sector of services. [online] Available:<br />

http://steconomice.uoradea.ro/anale/volume/2008/v4-management-marketing/085.pdf [September 5, 2009].<br />

Irani, Z. (2002) “Information systems evaluation: navigating through the problem domain”, Journal of Information &<br />

Management, Vol 40, Issue 1, pp 11-24.<br />

Jones, P., Beynon-Davies, P., and Muir, E. (2003) “E-business barriers to growth within the SME sector”, Journal of<br />

Systems and Information Technology, Vol 7, Issue 1/2, pp 1 – 25.<br />

Kollberg, M., and Dreyer, H. (2006) “Exploring the impact of ICT on integration in supply chain control: a research<br />

model”, Department of Production and Quality Engineering, Norwegian University of Science and Technology,<br />

Norway.<br />

Lawal, A. A. (2002) Entrepreneurship Development in Small and Medium Enterprises: Key Success Factor. Lagos<br />

Journal of Business, Vol 2, No 1, pp 35-46.<br />

Lawrence, J. E. (2008) “The challenges and utilization of E-commerce: use of Internet by Small to Medium Sized<br />

Enterprises in the United Kingdom”, Information, Society & Justice, Vol 1, No 2, pp 99-113.<br />

Leenders, M. A. A. M. and Wierenga, B. (2002) “The Effectiveness of Different Mechanisms for Integrating Marketing<br />

and R&D. Journal of Product Innovation Management, Vol 19, No 4, pp 305-317.<br />

Love, P. E. D., Irani, Z. and Edwards, D. J. (2004) “Industry-Centric benchmarking of information technology benefits,<br />

costs and risks for small and medium sized enterprises in construction”, Automation in Construction, Vol 13, No<br />

4, pp 507-524.<br />

Maldeni, H. M. C. M., and Jayasena, S. (2009) “Information and Communication Technology Usage and Bank<br />

Branch Performance”, The International Journal on Advances in ICT for Emerging Regions (ICTer), Vol 2, No 2,<br />

pp 29 – 37.<br />

Ojukwu, D. (2006) “Achieving sustainable growth through the adoption of integrated business and information<br />

solutions: a case study of Nigeria small and medium sized enterprises”, Journal of Information Technology<br />

Impact, Vol 6, No 1, pp 47-60.<br />

Ongori, H. and Migiro, S. O. (2010) “Information and Communication technology adoption: a literature review”,<br />

Journal of Chinese Entrepreneurship, Vol 2, No 1, pp 93-104.<br />

Orlikowski, W.J. and Baroudi, J. J. (1991) “Studying Information Technology in organisations: Research Approaches<br />

and Assumptions”, Information Systems Research, Vol 2, pp 1-28.<br />

490


Idisemi Apulu and Ann Latham<br />

Parida, V., Westerberg, M., and Yinenpaa, H. (2009) “How do small firms use ICT for business purpose? A study of<br />

Swedish technology-based firms”, International Journal of Electronic Business, Vol 7, No 5, pp 536-551.<br />

Pavic, S., Koh, S. C. L. Simpson., M. and Padmore, J. (2007) “Could e-business create a competitive advantage in<br />

UK SMEs?” Benchmarking: An International Journal, Vol14, No 3, pp 320-351.<br />

Pokharel, S. (2005) “Perceptions on information and communication technology perspectives in logistics: A study of<br />

transportation and warehouses sections in Singapore”, Journal of Enterprise Information Management, Vol 18,<br />

Issue 1/2, pp 136-149.<br />

Porter, M. E. and Millar, V.E. (1985) “How information gives you competitive advantage”, Harvard Business Review<br />

(July-August), Vol 63, Issue 4, pp 149-174.<br />

Remenyi, D., and Williams, B. (1996) “The Nature of Research: Qualitative or Quantitative, Narrative or<br />

Paradigmatic?”, Information Systems Journal, Vol 6, pp 131-146.<br />

Sarosa, S. and Underwood, J. (2005) “Factors affecting ICT adoption within Indonesian SMEs: Manager’s<br />

perspective”, PACIS 2005 proceedings, paper 6.<br />

Sauer, C. (1993) Why Information Systems Fail: A case Study Approach. Henley-on-Thames: Alfred Wailer.<br />

Siagala, M. (2003) “The information and communication technologies productivity impact on the UK hotel sector”,<br />

International Journal of Operations & Production Management, Vol 23, No 10, pp 1224-1245.<br />

Small and Medium Enterprises Agency of Nigeria (SMEDAN) (2005) “SME Success Digest”, Volume 3, No 1,<br />

January-June, 2005.<br />

Swift, M. (2009) “ICTs for SMEs – How ICT can be a force of viability”, ICT for SMEs Seminar, The Cascadia Hotel &<br />

<strong>Conference</strong> Centre Port of Spain, Trinidad & Tobago Wednesday, 25th March, 2009. [online]<br />

http://www.iberpymeonline.org/Trinidad0309/MichaelSwift.pdf [Accessed 10th June 2009].<br />

Van Akkeren, J and Cavaye, L. M. A. (1999) “Factors affecting entry-level internet technologies adoption by small<br />

business in Australian: an empirical study”, 10 th Australasian <strong>Conference</strong> on Information Systems, Vol 199, pp<br />

1071-1083.<br />

Wielicki, T. and Arendt, L (2010) “A knowledge-driven shift in perception of ICT implementation barriers: comparative<br />

study of US and <strong>European</strong> SMEs”, Journal of Information Science, Vol 36, No 2, pp 162-124.<br />

Yin, R. K (2003) Case Study Research: Design and Methods. 3rd Edition, Applied Social Science Method Series,<br />

Vol. 5, Sage Publications, Thousand Oaks, CA.<br />

Zhouying, S., Weidong, L., and Turong, M. (2009) “The sustainable development of ICTs in China: An empirical<br />

study of small and medium enterprise”, Chinese Journal of Population, Resources and Environment, Vol 7, No<br />

3, pp 20-29.<br />

491


Establishing the Suitability of Dynamic Web Applications Development<br />

Platforms for Learning web Techniques in Tertiary<br />

Institutions<br />

Johnson Dehinbo<br />

Tshwane University of Technology, Pretoria, South Africa<br />

DehinboOJ@tut.ac.za<br />

Abstract: The availability of many platforms for developing dynamic Web applications suggests the existence of the<br />

problem of choosing the most suitable platform for learning the concepts of web applications development by undergraduate<br />

students in tertiary institutions. Students may not perform at their best capacity level if the platform chosen<br />

by the institution is not very suitable for learning the relevant concepts. As part of the framework to determine the<br />

most suitable platform for leaning web applications development in tertiary institutions, this study establishes a set of<br />

criteria for evaluating the suitability for learning the concepts of Web techniques in Web application development<br />

such as Web page serving, HyperText Transmission Protocol (HTTP) basics, Server information, Form validation<br />

and processing, as well as maintenance of Session and Application states. These criteria were applied by evaluating<br />

four platforms namely Java Servlets, Java Server Pages, Active Server Pages and PHP using various research<br />

methods including descriptive inquiry, document analysis, observations and programming tests. While ASP was<br />

found to be more suitable on applying the criteria, the significance of the study lies in the establishment of a comprehensive<br />

but specific set of criteria that can be used as a scientific basis for selection.<br />

Keywords: web applications development platforms, programming languages, web techniques, comparison frameworks<br />

1. Background of the study<br />

This study establishes part of a framework containing various criteria that can be used to evaluate dynamic<br />

Web platforms to determine a suitable platform for learning Web applications development in tertiary<br />

institutions. This part of the framework is to determine platform that will be suitable for learning the<br />

concepts of Web techniques in Web applications development.<br />

There are currently many platforms for implementing dynamic application programs on the World Wide<br />

Web and for learning dynamic web applications development in tertiary institutions today. Lim (2002)<br />

states that information systems departments need to reexamine their curricula in order to prepare students<br />

to face the challenge of being productive in a computing world swamped with web technologies.<br />

The use of various programming and scripting languages could lead to elements of repetition and confusion.<br />

Yet, comprehension is very important in undergraduate studies (Wiedenbeck 1999:5).<br />

The comprehension of relevant concepts by students may be affected by the choice of platform because<br />

according to Sebesta (1996:2-3), the language in which programmers develop software places various<br />

limits on the development efforts. Thus, the choice of dynamic web application development platform<br />

should involve critical evaluation in line with the statement below:<br />

Web development tools need to be analyzed in terms of its purpose (what it is designed to<br />

do), technology (ease of use, robustness, scalability, security, performance, etc.), support<br />

(portability, cost, ISP support), and how well it works in the real world (Ashenfelter 1999:<br />

105).<br />

With regard to the purpose-based analysis of the platforms, appropriate Web techniques are critical for<br />

developing dynamic Web applications. The research question then is: How do we determine the dynamic<br />

Web platform that will be suitable for learning the concepts of Web techniques in Web applications development<br />

by undergraduate students?<br />

The objectives of the study are therefore given as follows:<br />

1.) To identify the concepts of Web techniques that needs to be learnt by undergraduate students of<br />

Web applications development in tertiary institutions.<br />

2.) To establish the criteria to determine suitability for learning the concepts.<br />

3.) To apply the criteria by using them to evaluate the suitability of four specified platforms.<br />

492


2. Literature review<br />

Johnson Dehinbo<br />

Evaluating programming languages, development platforms and tools is very important for understanding<br />

the effects on novice programmers, but is difficult to carry out (Wiedenbeck et al. 1999). This could be the<br />

reasons why Apte et al. (2003) note that a study of existing literature showed varying conclusions about<br />

the superiority of one dynamic Web platform over another. Prechelt (2000) also indicates that programmers<br />

usually hold varied strong opinions on development platforms. These are illustrated in the various<br />

approaches towards the comparisons of platforms in the studies reviewed below.<br />

2.1 Previous studies comparing platforms<br />

In a survey of middleware platforms, Cooper (2001) concluded that ColdFusion is fast to learn and fast to<br />

use, Common Gateway Interface (CGI) is also fast to learn and that Servlets are hard to learn and use,<br />

even by someone who already knows Java. Bishop and Hurter (1999) examined the Scripting languages;<br />

Tcl/Tk, Perl and Python and found out that Python incorporated the features of Modula-3 into its scripting<br />

syntax thus making it suitable for "programming in the large". Programs in Python were also found to be<br />

typically much shorter than equivalent in C or Java.<br />

Prechelt (2000) observes that designing and writing programs in the scripting languages, namely Perl,<br />

Python, Rexx, or Tcl takes no more than half as much time as writing it in C, C++, or Java. Moreover, the<br />

resulting program is only half as long. Kruse (2003) illustrates the differences in the strengths and weaknesses<br />

of Personal Home Page (PHP) and Java.<br />

However, most of these studies did not use any criterion as a basis for the comparisons. The comparisons<br />

in these studies seem to be based on intuition rather than scientific facts. Comparison needs to be<br />

based on a variety of factors supported by scientific facts and results.<br />

A common criterion used in other studies is performance. Renaud et al. (2003) indicate various metrics<br />

can be used to measure performance. These include response or waiting time, synch delay, number of<br />

messages exchanged, throughput, communication delay, node fairness, Processor and memory usage.<br />

Marshak and Levy (2003) evaluated platforms only in terms of user-perceived latency.<br />

Interestingly, Vinoski (2003) pointed out that performance has somehow been overemphasized as people<br />

check only those qualities that are easily measurable, such as performance. This has unintentionally led<br />

many programming language users to presume that “high performance” is the same as “high quality”.<br />

Such presumptions could be entirely meaningless, depending on the nature of one’s application. So,<br />

comparisons should involve other relevant factors as criteria.<br />

2.2 Studies using various criteria in their comparisons<br />

Cecchet et al (2003) evaluate three specific mechanisms namely PHP, Java Servlets, and Enterprise<br />

Java Beans (EJB) with respect to performance and ease of development. The study attributes PHP’s better<br />

performance to the fact that it executes as a module in the Web server, sharing the same process<br />

(address space), thereby minimizing communication overhead. This is unlike Java Servlets which run in a<br />

Java Virtual Machine (JVM) as a separate process from the Web server (Cecchet et al 2003). In terms of<br />

ease of development, Cecchet et al (2003) explain that PHP scripts are easy to write because they embed<br />

code directly into an HTML page.<br />

In examining ASP, PHP and ASP.NET, Hartman (2001) mentioned two factors that complicate choosing<br />

a scripting environment. First, there is the issue of culture among developers which has a lot to do with<br />

the ideological camps to which they belong. If they love to tinker with source code to develop solutions<br />

more efficient than off-the-shelf products, and embellished their cubicles with defaced portraits of Bill<br />

Gates, they would probably prefer to use PHP. Convenience lovers would probably prefer to use ASP.<br />

Yet, very few developers are equally willing to use both, or talk about "the other" technology without a<br />

trace of disdain. Hartman’s second factor is the website's future scalability and functional requirements<br />

which might restrict which servers and platforms the site could run on or impact the feasibility of developing<br />

future features, such as database-linked connectivity (Hartman 2001).<br />

However, the most important criterion is the suitability for doing the job for which a tool is needed. This<br />

view is presented next.<br />

493


Johnson Dehinbo<br />

2.3 Studies emphasizing suitability for achieving the purpose of systems<br />

Web development tools need to be analyzed in terms of its purpose (Ashenfelter 1999:105). Similarly, for<br />

choosing a language or platform that is suitable for learning introductory programming, Holt et al. (1997)<br />

use the criteria that it should be appropriate for introducing programming concepts used in the real world<br />

such as in business, science and government; and it should encourage systematic problem solving. Similarly,<br />

Kolling and Rosenberg (1996) suggest clean concepts directly reflecting the theoretical model and<br />

avoidance of conceptual redundancy as achieving the same thing in a variety of ways can mean flexibility<br />

to the expert, but usually confuses beginners.<br />

Moreover, Hadjerrouit (1998) evaluates the suitability of Java as a first programming language using the<br />

criteria like: programming concepts to be taught including problem solving skills, algorithmic thinking and<br />

structured programming; use in subsequent courses and programming paradigm support.<br />

These studies emphasize the need for the suitability for doing the job for which a tool is needed as Web<br />

techniques is required in Web applications development.<br />

2.4 Summary of the literature review<br />

Table 1 summarizes the literature review. A key point observed is that choosing a suitable platform<br />

should involve exhaustive evaluations of various options based on various relevant criteria that are<br />

backed by scientific facts and results. This study fills this gap in the body of knowledge by being unique in<br />

the following ways: The advantages and strengths of platforms should be examined and ranked in the<br />

light of certain desired qualities relevant to specific use such as the suitability for learning Web techniques,<br />

just as for example, the ease of learning increases productivity in institutions than flexibility which<br />

could be preferable in industries. This informs the identification of desirable concepts of Web techniques<br />

before establishing criteria for evaluations.<br />

Table 1: Scoring for the platforms based on the criteria on web techniques.<br />

Sources Tasks / Conclusions Limitations / Points noted<br />

Cooper(2001) Surveyed some middleware platforms and concludes<br />

that Coldfussion is fast to learn and use while Servlet is<br />

hard to learn and use.<br />

Bishop & Hurter<br />

(1999)<br />

Examined scripting languages Tcl/Tk, Perl and Python<br />

and concludes that Python programs are typically<br />

shorter.<br />

Prechelt(2000) Observes that designing and writing programs in the<br />

scripting languages, namely Perl, Python, Rexx, or Tcl<br />

takes no more than half as much time as writing it in C,<br />

C++, or Java. Moreover, the resulting program is only<br />

half as long<br />

Kruse (2003) Illustrates the differences in the strengths and weaknesses<br />

of Personal Home Page (PHP) and Java<br />

Renaud et al.<br />

(2003)<br />

Marshak and<br />

Levy (2003)<br />

Indicate various metrics can be used to measure performance.<br />

These include response or waiting time,<br />

synch delay, number of messages exchanged,<br />

throughput, communication delay, node fairness, Proc-<br />

essor and memory usage<br />

Also evaluated platforms only in terms of userperceived<br />

latency.<br />

Vinoski (2003) Pointed out that performance has somehow been<br />

overemphasized as people check only those qualities<br />

that are easily measurable, such as performance<br />

494<br />

Systematic approach not<br />

used in evaluating ease<br />

of learning and ease of<br />

use<br />

Comparison not based<br />

on specific use<br />

Did not use any specific<br />

criterion as a basis for<br />

the comparisons.<br />

The comparisons seem<br />

to be based on intuition<br />

rather than scientific<br />

facts<br />

Comparison concentrates<br />

on performance<br />

which Vinoski states has<br />

been over-emphasized<br />

Comparison based on<br />

latency form of perform-<br />

ance as above<br />

Performance may not be<br />

the ultimate depending<br />

on the nature of one’s<br />

application


Cecchet et al<br />

(2003)<br />

Johnson Dehinbo<br />

Evaluate PHP, Java Servlets, and Enterprise Java<br />

Beans (EJB) with respect to performance and ease of<br />

development. The study attributes PHP’s better performance<br />

to the fact that it executes as a module in the<br />

Web server, sharing the same process (address<br />

space), thereby minimizing communication overhead.<br />

Serve as example towards<br />

multiple factors<br />

used in the comparisons<br />

Sources Tasks / Conclusions Limitations / Points noted<br />

Hartman (2001) Examined ASP, PHP and ASP.NET, mentioned two<br />

factors that complicate choosing a scripting environment.<br />

These are the issue of culture among developers<br />

dealing with the ideological camps to which they belong<br />

and the website's future scalability with functional<br />

Ashenfelter<br />

(1999: 105)<br />

requirements.<br />

Established that “Web development tools need to be<br />

analyzed in terms of its purpose (what it is designed to<br />

do), technology (ease of use, robustness, scalability,<br />

security, performance, etc.), support (portability, cost,<br />

ISP support), and how well it works in the real world”<br />

Holt et al. (1997) Use the criteria that it should be appropriate for introducing<br />

programming concepts used in the real world<br />

such as in business, science and government; and it<br />

Kolling and<br />

Rosenberg<br />

(1996)<br />

Hadjerrouit<br />

(1998)<br />

should encourage systematic problem solving<br />

Suggest clean concepts directly reflecting the theoretical<br />

model and avoidance of conceptual redundancy as<br />

achieving the same thing in a variety of ways can<br />

mean flexibility to the expert, but usually confuses be-<br />

ginners<br />

Evaluates the suitability of Java as a first programming<br />

language using the criteria like: programming concepts<br />

to be taught including problem solving skills, algorithmic<br />

thinking and structured programming; use in<br />

subsequent courses and programming paradigm support.<br />

3. Research design and methodology<br />

Also serve as example<br />

towards multiple factors<br />

used in the comparisons<br />

This serves as for this<br />

study’s evaluation based<br />

on ‘purpose’ which is the<br />

learning of Web tech-<br />

niques<br />

This serves as a pointer<br />

into the formulation of<br />

criteria to be established<br />

Also serves as a pointer<br />

into the formulation of<br />

criteria to be established<br />

Emphasize the need for<br />

the suitability for doing<br />

the job for which a tool is<br />

needed.<br />

An interpretive research design is used involving elements of descriptive, analytical and comparative<br />

studies. McMillan and Schumacher (2001:33) state that while a descriptive study describes a system with<br />

the aim of characterizing it as it is, by using numbers, comparative study investigates the differences,<br />

thereby taking descriptive study a step further. Therefore, a descriptive approach is used to characterize<br />

desirable features of the platforms towards establishing a set of criteria. In applying the established criteria,<br />

an analytical approach is used to analyze and evaluate platforms to determine the level of satisfaction<br />

of the criteria for learning Web techniques.<br />

3.1 Research methods adopted to obtain the results<br />

The descriptive method used involves document review and study of various manuals and textbooks for<br />

different platforms from various sources and established body of knowledge, including those written by<br />

the designers and originators of the platforms as well as various websites for the applicable web servers<br />

such as IIS and tomcat, to identify features that could enhance the learning of Web techniques in Web<br />

applications development. These features were characterized to establish criteria for evaluating the suitability<br />

of platforms for learning Web techniques. The established criteria serve as a model. Bowling<br />

(2002:141) describes models as abstract representations of the essential characteristics of phenomena<br />

of interest that make explicit, the relationships and comparison between the characteristics. The model is<br />

then applied for analyzing, evaluating and comparing specific platforms.<br />

The analytical method is aimed at evaluating specific platforms to ascertain their level of satisfaction of<br />

the established criteria. Therefore, documents were reviewed for the specific platforms to be evaluated<br />

and compared, and answers were sought to the questions and the availability of features that serve as<br />

the criteria.<br />

495


Johnson Dehinbo<br />

Moreover, experience was used to physically observe the support for various Web techniques by the platforms.<br />

Bowling (2002:358) describes observation as a research method in which the investigator systematically<br />

watches, listens to and records the phenomenon of interest. Scores were then assigned to each<br />

platform based on the availability of desired features, the level of support enabled on specific tasks or the<br />

inherent characteristics of the platform.<br />

3.2 Measuring scale used<br />

Using close-ended "Yes/No" questions, the measuring tool has values on a scale of 1 to 3, where:<br />

3 = "Yes",<br />

2 = "Not quite or with some workaround", and<br />

1= "No".<br />

A scale of 1 to 3 avoids subjective situations where it could be difficult to distinguish between, for example,<br />

a score of 3 or 4 in a scale of 1 to 5.<br />

3.3 Establishing reliability and validity of the descriptive inquiry<br />

It is important that a measuring scale or instrument be consistent and reliable. It should produce more or<br />

less the same accurate results every time it is applied (Coertze & Heath 1997: 78). Also, Coertze and<br />

Heath (1997:79) indicate that validity is concerned with soundness or the effectiveness of the measuring<br />

instrument. As a way of increasing validity, answers to the criteria questions were sought from established<br />

and recognized sources. The accompanying references are provided for verification or for more<br />

information. This is supplemented with practical experiences and program tests confirming the satisfaction<br />

of some of the criteria established.<br />

Also, to increase reliability, the quantitative characterization and evaluation using numbers would enhance<br />

clarity in the choice of platform with the highest score. This is unlike just using qualitative sentences<br />

to evaluate the platforms, at the end of which it could be difficult to say which platform is really<br />

more suitable. Furthermore, reliability is enhanced with the scale of 1 and 3 thereby avoiding subjective<br />

situations of distinguish between, say, a score of 3 or 4 in a scale of 1 to 5.<br />

3.4 Data analysis<br />

The data analysis was done using simple statistical parametric analysis, such as sums and means. The<br />

scores for all the criteria were summed up for each web based dynamic platform to obtain a total score<br />

from which the platform with the highest overall score can be identified.<br />

3.5 Limitations<br />

As an interpretive study, a particular limitation is that of the subjective view of the researcher. However,<br />

users can add/remove certain criteria or change their weights.<br />

4. Development of the criteria for the evaluations<br />

It is important that a suitable platform satisfies the purpose for which it is needed which in this case is the<br />

learning of Web techniques. Therefore, important Web techniques are first identified before establishing<br />

desired qualities and questions that serve as criteria for evaluating the suitability for implementing and<br />

learning these concepts.<br />

4.1 Important web techniques<br />

Web applications development is primarily about creating dynamic Web applications. However, dynamic<br />

contents are usually surrounded by static contents. This necessitates learning the generation of both<br />

static and dynamic contents. This involves techniques such as Web page serving, HTTP Basics, Server<br />

information, Form processing, as well as maintenance of Session and Application states.<br />

Web page serving<br />

This involves the creation, loading and redirection of Web pages. Students need to learn:<br />

496


Johnson Dehinbo<br />

How to install “normal HTML codes” on the Web server? Which directory the “HTML files” will be<br />

stored in order for the Web page to be accessible with appropriate URL addresses?<br />

How to access Web pages from the browsers? How the appropriate folders/directories are mapped<br />

relative to the default Web root directory. How appropriate URLs are mapped to the installed application/HTML<br />

files in relative folders?<br />

How to redirect Web pages to other pages? How pages are linked in HTML and how Web forms call<br />

the necessary programs to preprocess the form data?<br />

HTTP basics<br />

The Web runs on HTTP which governs how Web browsers request files from Web servers and how the<br />

servers send the files back. A basic understanding of HTTP is necessary to understand the various Web<br />

techniques. This includes knowledge of the HTTP request messages; its content, which includes some<br />

header information, and optional body; the Web server’s response message including some header information<br />

and usually a body (Lerdorf & Tatroe 2002:158).<br />

Server information<br />

Much information about the Web server and the client will usually be made available to the Web applications<br />

for effective Web management. Although some server information are obtainable from the response<br />

header, students need to learn how further information can be obtained from the server. Such further information,<br />

according to Lerdorf and Tatroe (2002:160-161) includes the hostname or IP address, the<br />

server port number, the fetch method (GET/POST), the path information (e.g.<br />

“…/students/science_faculty”), the script/application/program name useful for self-referencing scripts, the<br />

query string in the URL (e.g., “name=Johnson+age=38”), the client PC’s name/IP_address, the authentication<br />

method used, etc.<br />

Form processing<br />

Using form data moves data from a Web page to a server-side program (Hall, 1999). Students need to<br />

learn many tricks and techniques for working with forms:<br />

Attaching information as a query string at the end of the URL after a question mark, for GET requests,<br />

or sending it on a separate line, for POST requests (Sun Microsystems 1997).<br />

The significance of using POST rather than GET, with respect to the security of the transmitted data.<br />

Techniques of implementing self-processing pages that both generate the form and process it.<br />

Methods of extracting the sent information from form data which differs for various Web applications<br />

platforms, along with the ways of converting one type of data to another.<br />

File Uploading and how to handle the uploaded files?<br />

Form Validation<br />

It is necessary to validate entered data before using or storing it for later use. Several strategies are<br />

available for validating data. One involves using JavaScript or VBScript on the client side. However, since<br />

the user can choose to turn JavaScript or VBScript off, or may even be using a browser that does not<br />

support it (Lerdorf & Tatroe 2002:173), a more secure choice is to use the dynamic Web platform to do<br />

the validation.<br />

Students therefore need to learn how to:<br />

Check that a value was supplied: i.e. form variables are non-empty.<br />

Check that an email address is valid by requiring an “@” sign, checking for the presence of period “.”<br />

in the domain name, checking the last two or three digits of the domain name against the country<br />

codes and organization domains respectively.<br />

Check that a supplied filename is local and exists. This can be ensured by allowing the “browsing” of<br />

the filename on the required media.<br />

Validate a field to ensure that it contains a nonnegative integer.<br />

Ensure correctness of the comparison of a form value with other values by trimming leading/trailing<br />

spaces, converting to lowercase/uppercase etc.<br />

497


Maintaining Session and Application states<br />

Johnson Dehinbo<br />

HTTP is a stateless protocol (Lerdorf & Tatroe 2002:173; Sun Microsystems, 2002, Module 8: 3), which<br />

means that once a Web server completes a client’s request for a Web page, the connection between the<br />

two ceases to exist. Thus, there is no way for a server to recognize that a sequence of requests all originate<br />

from the same client.<br />

State is useful though. To build a shopping cart application for example, it is necessary to keep track of a<br />

sequence of requests from a single user (Lerdorf & Tatroe 2002:178). According to Hall (1999), maintaining<br />

states can help the server to provide visitors with a number of conveniences: Firstly, it enables identifying<br />

a user during an e-commerce session. The user selects and adds items to the shopping cart, and<br />

then continues shopping. Since the HTTP connection is closed after each page is sent, how does the<br />

store know that the user is the same user who put the previous item in his cart? Maintaining states using<br />

various techniques is a way of accomplishing this.<br />

Secondly, it is inconvenient to remember the username and password for many sites. An alternative is for<br />

the server to give a unique user-ID, and when the client reconnects at a later date, the user-ID is returned,<br />

which is verified by the server. Thirdly, it allows customizing a site and lastly, it enables the focusing<br />

of advertisement for displaying "directed" advertisements than "random" ones. This is possible if the<br />

server “remembers” clients’ interests.<br />

Techniques are available to get around the Web’s lack of state, and to keep track of state information<br />

between requests (i.e. session tracking). These include using Hidden form fields, Cookies, and URL rewriting.<br />

Hidden form fields can be used to pass around information using unique identifier assigned. With<br />

URL rewriting, where every local URL on which the user might click is dynamically modified to include<br />

extra information (Lerdorf & Tatroe 2002:178) such as a user-identification or session ID. A cookie is a bit<br />

of information that the server can give to the client. On every subsequent request, the client will give that<br />

information back to the server, thus identifying the client (Hall 1999: Lerdorf & Tatroe 2002:179). In using<br />

cookies, students will need to learn: How to retrieve the Session object from the request object passed?;<br />

How to store the Session attributes?; Accessing the Session attributes?; Destroying the Session; Tedious<br />

details, problems and limitations of Session management.<br />

Session and Application object store state information for use by script files and can also be used to<br />

share information among all users of a given application. One can therefore use Application variables to<br />

store global information, such that different users can modify Application variables simultaneously (MSDN<br />

2002: Module 5: 18). Students will need to learn how to create, lock, store and unlock the Application objects.<br />

4.2 Criteria for evaluating the suitability for teaching Web techniques<br />

From the concepts above, this section now establishes the criteria to ensure that Web applications development<br />

platforms could handle Web techniques with the following features:<br />

Facility for static Web page serving by the associated Web server: this is necessary to illustrate the<br />

use of the platform for static Web contents.<br />

Facilities for extracting information from the HTTP request header: this indicates the use of the request<br />

header as a medium for conveying useful information.<br />

Facilities to embed or set information into the HTTP response header: this illustrates the use of the<br />

response header to convey useful information.<br />

Provision of other relevant information by the Web server: this illustrates the wealth of information<br />

available through the server.<br />

Availability of default folders for placing program codes: these will be easier for beginners to use before<br />

they have mastered the art of mapping to various directories.<br />

Possibilities of sending information via the service method that could implement both GET and POST<br />

methods: this will ensure that the rest of the codes are not affected by changes between GET and<br />

POST methods.<br />

Facilities for form validation, e.g. checking for non empty fields, range of values, content of fields, and<br />

using hidden variables to check if form has been filled once, etc.; this will illustrate the importance of<br />

form validations.<br />

498


Johnson Dehinbo<br />

Facilities for identifying users in a session, e.g. using cookies, hidden variables or URL rewriting to<br />

illustrate the techniques used to overcome HTTP’s statelessness.<br />

Possibility of setting “age” for cookies thereby making cookies last longer or alternatively disallowing<br />

cache of pages.<br />

Facilities for maintaining Application objects: this illustrates the techniques for maintaining global Application<br />

states.<br />

5. Summary of the results of applying the criteria<br />

Application of the established criteria involves evaluating the suitability of four platforms namely Java<br />

Servlets, JSP, ASP and PHP for learning the Web techniques. Table 2 below gives the scoring for the<br />

platforms based on the criteria. From the table, it is obvious that most of the Web techniques features are<br />

supported by all the platforms, possibly because that is the major goal of all platforms. In order to serve<br />

static information that accompanies dynamic contents, the associated Web servers all have facilities for<br />

static Web page serving.<br />

The HTTP headers serve to pass valuable information, including error status, and as such all the platforms<br />

have facilities to extract information from the HTTP request-headers, and to embed information into<br />

the HTTP response-headers (Lerdorf & Tatroe 2002:175-176). All the platforms have methods of obtaining<br />

relevant information about client and server.<br />

Table 2: Scoring for the platforms based on the criteria on Web techniques.<br />

Criteria questions Servlet JSP ASP PHP<br />

1 Are there facilities for static Web page serving by the associated<br />

server?<br />

3 3 3 3<br />

2 Are there facilities to extract information from the HTTP request<br />

headers?<br />

3 3 3 3<br />

3 Are there facilities to embed or set information into the HTTP re- 3 3 3 3<br />

sponse headers?<br />

4 Are there ways of obtaining relevant information from the server? 3 3 3 3<br />

5 Is there availability of default folders for placing program codes? 3 3 3 3<br />

6 Is there the possibility of sending information via a service method<br />

that could implement both GET and POST methods?<br />

7 Do numeric values accepted from a Web form retain their numeric<br />

type instead of being in string format which still has to be recon-<br />

verted ?<br />

8 Are there simple facilities or functions for form validation, e.g.<br />

checking non-empty fields, range of values, field contents etc.?<br />

9 Are there facilities for identifying users in a session, e.g. using<br />

cookies, hidden variables or URL rewriting?<br />

10 Is there the possibility of setting the age of cookies thereby making<br />

cookies last longer?<br />

11 Are there facilities for maintaining application objects so as to maintain<br />

global application states?<br />

12 Pedagogically, is the platform good for teaching and learning Web<br />

techniques?<br />

TOTAL<br />

Scale: 3 = "Yes", 2 = "Not quite or with some workaround", and 1= "No".<br />

3<br />

1<br />

2<br />

3<br />

1<br />

2<br />

3<br />

3<br />

3<br />

3<br />

3<br />

2<br />

3 3 3 3<br />

3 3 3 3<br />

Novice Web developers would prefer putting their codes in default folders rather than creating and mapping<br />

new folders. All the platforms have default folders for this purpose.<br />

Experience shows that passing information via GET methods embeds the values along with the URL and<br />

as such is useful for debugging. But for good security, POST does not show the information on the URL.<br />

All the platforms allow for the sending of information via a service method that could implement both GET<br />

and POST methods.<br />

499<br />

2<br />

2<br />

31<br />

2<br />

2<br />

31<br />

3<br />

3<br />

36<br />

2<br />

3<br />

34


Johnson Dehinbo<br />

When numeric values are accepted from a Web form, it is better that they retain their numeric type instead<br />

of being in String format which still has to be reconverted. This is true in all the platforms except<br />

Java-based ones where form values that are not Strings must be parsed because they are all passed as<br />

Strings (Sebesta 1996:435). One has to use methods such as parseInt() to convert back to integer (Wigglesworth<br />

2000:821). However, experience has shown that in the case of ASP, when the operator used<br />

on the variables is an addition operator “+”, it always assumes the String concatenation operator “+”,<br />

thereby giving results such as “2+5 = 25”. ASP similarly provides the conversion functions such as CInt()<br />

to explicitly convert to integer (Deitel et al. 2001: 741) in such situations.<br />

The fact that the need for the explicit conversion explained above arising due to confusion between addition<br />

and String concatenation is not necessarily applicable in the case of other arithmetic operators such<br />

as subtraction “-”, multiplication “*” and division “/” makes such problems difficult to identify. If experienced<br />

programmers have difficulties in identifying such problems, one can imagine how long it will take<br />

students to do the same.<br />

Validation of values accepted from Web forms is important to ensure the integrity of the processing. In<br />

the Java-based platforms, one can use the length() method to determine whether the field is empty and<br />

one can “catch the NumberFormatException” if the entered value is non-numeric (Wigglesworth 2000:<br />

821). ASP has simple functions such as IsDate(), IsNumeric(), and IsEmpty() to assist in the validation<br />

(Deitel et al. 2001: 735). Also, there are facilities or functions for form validation such as checking nonempty<br />

fields, range of values, content of fields etc. in the PHP platform (Lerdorf & Tatroe 2002:173), but<br />

not as simple as in ASP.<br />

To remedy HTTP’s statelessness, all the platforms have facilities for identifying users in a session via<br />

cookies, hidden variables or URL rewriting, and there is the possibility for setting the age of cookies to<br />

make them last longer (Lerdorf & Tatroe 2002:178-179). In addition, all the platforms have ways of maintaining<br />

global Application states. Pedagogically, ASP and PHP, from experience, seem to be more suitable<br />

for teaching and learning Web techniques due to the simplicity of their commands and functions. In<br />

summary, applying the established criteria reveals that ASP has the highest score followed by PHP and<br />

then the Java-based platforms. It should however be noted that the scores are subjectively based on author’s<br />

knowledge and experience as well as the current design of the platforms. They are thus subject to<br />

change. Therefore, the emphasis of this study is on the establishment of the criteria that enables the selection<br />

among the platforms. Users can thus adapt these criteria to their own taste and can also test or<br />

apply the criteria on other platforms.<br />

6. Conclusions<br />

It is important to ensure that a platform selected for learning Web applications development is suitable for<br />

implementing and learning Web techniques. This is because Web development platforms differ from the<br />

general programming languages in their support for Web techniques used for generating both static and<br />

dynamic Web contents. By studying the features of many platforms, features that are desirable for supporting<br />

the Web techniques in Web applications are identified. These enabled the establishment of criteria<br />

to determine the suitability of the platforms for implementing and learning Web techniques. While ASP<br />

was found to be slightly more suitable on applying the criteria, the significance of the study lies in the establishment<br />

of a comprehensive but specific set of criteria that can be used as a scientific basis for selection.<br />

References<br />

Apte, V., Hansen, T. and Reeser, P. (2003) “Performance comparison of dynamic Web platforms.” Computer Communications.<br />

26 (8) pp. 888 – 898.<br />

Ashenfelter, J.P. 1999. Choosing a Database for Your Web Site, USA : Wiley Computer Publishing.<br />

Bishop, J. and Hurter, R. 1999. Competitors to Java: Scripting languages. In: Proceedings of the South African<br />

Computer Lecturers Association (SACLA) conference. June 1999. Golden Gate, South Africa. [Online]. Available<br />

WWW: http://www.cs.up.ac.za (Accessed 12/03/2009).<br />

Bowling, A. 2002. Research methods in health. 2 nd edition. UK: Open University Press.<br />

Cecchet, E., Chanda, A., Elnikety, S., Marguerite, J. & Zwaenepoel, W. 2003. Performance Comparison of Middleware<br />

Architectures for Generating Dynamic Web Content. Lecture Notes in Computer Science: Middleware.<br />

2672 (2003): 242-261.<br />

Coertze, D. and Heath, R. 1997. Research Methodology for Technikon Students: A practical Approach. South Africa :<br />

Technikon Natal publishing.<br />

Cooper, R. 2001. Software for managing Web sites. In: Proceedings of the South African Institute of Computer Scientists<br />

and Information Technologists (SAICSIT) annual conference. September 2001. Pretoria: South Africa.<br />

500


Johnson Dehinbo<br />

Deitel, H.M., Deitel, P.J. and Nieto, T.R. (2001) e-Business & e-Commerce: how to program. Prentice Hall, New Jersey.<br />

Hall, M. 1999. Tutorial on Servlet and JSP [Online]. Available from: http://www.apl.jhu.edu/~hall/java/Servlet-Tutorial/<br />

[Accessed: 5/06/2009].<br />

Hartman, H. 2001. Tools for dynamic Web sites: ASP vs PHP vs ASP.NET, Seybold Report Analysing Publishing<br />

Technologies, 15339211. 1 (12).<br />

Holt, R., Wortman, D., Barnard, D. and Cordy, J.R. (1977) “SP/k: a system for teaching computer programming.”<br />

Communications of the ACM. 20 (5) pp. 301-309.<br />

Kolling, M. and Rosenberg, A. (1996) “Blue: A language for teaching object-oriented programming.” Proceedings of<br />

the 27th SIGCSE Technical Symposium on Computer Science Education, Philadelphia. New York. pp.190-194.<br />

Kruse, W. (2003) “A comparison of PHP and J2EE.” Technologiae. 1 (2003) pp. 110-117.<br />

Lerdorf, R. and Tatroe, K. 2002. Programming PHP: Creating Dynamic Web Pages. USA : O’Reilly & Associates Inc.<br />

Lim, B.L. 2002. Teaching web development technologies: Past, present, and (near) future. Journal of Information<br />

Systems Education. 13 (2): 117-123.<br />

Marshak, M. and Levy, H. 2003. Evaluating web user perceived latency using server side measurements. Computer<br />

Communications. 26 (8): 872-887.<br />

McMillan, J.H. and Schumacher, S. 2001. Research in education. 5 th edition. USA: Addison Wesley Longman, Inc.<br />

MSDN. 2000. Introduction to Web Development Technologies. Course number 1912 Workbook. Part Number: X05-<br />

91011.<br />

Prechelt, L. 2000. An Empirical Comparison of Seven Programming Languages. Computer. 33 (10): 23-29.<br />

Renaud, K., Lo, J., Bishop, J., Van Zyl, P and Worrall, B. 1999. Algon: A framework for supporting comparison of<br />

distributed algorithm performance, In: Proceedings for PNDP ‘03 conference, February 2003. Genoa: Italy.<br />

Sebesta, R.W. 1996. Concepts of Programming Languages, 3 rd edition. USA: Addison-Wesley Publishing Company.<br />

Sun Educational Services. (2002) “Web Component Development With Java Technology.” SL-314 Student Guide.<br />

Revision A.1, Sun Microsystems, New York.<br />

Sun Microsystems. 1997. JDBC Guide: Getting Started [Online]. Available from:<br />

http://www.pearson.com.mx/component/component_java_21_days/jdk/1.2/doc/html/guide/jdbc/getstart/introTO<br />

C.doc.html [Accessed: 11/08/2008].<br />

Vinoski, S. (2003). “The performance presumption.” IEEE Internet Computing. 7 (2) pp. 88-90.<br />

Wiedenbeck, S., Ramalingam, V., Sarasamma, S. and Corritore, C.L. 1999. A comparison of the comprehension of<br />

object-oriented and procedural programs by novice programmers. Interacting with Computers. 11 (3): 252-282.<br />

Wigglesworth, J. and Lumby, P. 2000. Java Programming Advanced Topics. U.K.: Thompson Learning Course<br />

Technology.<br />

501


Human Resources Transformation Beyond Boundaries in<br />

Outsourcing Business Models - Expatriate Benchmarking<br />

Swathi Duppada 1 and Rama Chandra Aryasri 2<br />

1<br />

Satyam Computer Services Limited (Mahindra Satyam), Canada<br />

2<br />

School of Management Studies, JNTU, Hyderabad, India<br />

gsrkmedu@gmail.com,<br />

aryasri@yahoo.com<br />

Abstract: Human Resource (HR) divisions of multi national companies (MNCs) are under tremendous pressure<br />

globally with the challenges and opportunities with the outsourcing business models to maintain competitive position<br />

in the marketplace. Attracting the mobile talent with multi dimensional skill set to address effective, efficient and<br />

controllable business needs is becoming complex, hence expatriate management and training has gained much<br />

attention. Successful Expatriate assignments drive revenue, value and growth to the organization. The expatriation<br />

process requires huge amount of effort for analysis, planning, selection, training before departure of the associate to<br />

host country. The authors would like to bring the practical approaches that need to be considered in global business<br />

outsourcing model considering 3 dimensions – Associate delight, Customer delight and Investor delight with<br />

expatriate benchmarking. The research study also brings the expatriate management strategies in 3 categories –<br />

Onsite (foreign location / host country), Offshore (home country) and Near-shore (country close to host country, but<br />

with lesser delivery cost). To substantiate the research, the data is collected from several Human resource leaders<br />

and managers at various levels – HR Executives, HR Analysts, HR Managers, Senior Managers from IT<br />

organizations in different geographies through interviews and web based surveys. Statistical analysis are conducted<br />

on the data collected and the data analysis first revealed that 1) the organizations with good global management<br />

strategy had larger number of associates with better expatriate experience, steadier focus on leadership and resulted<br />

in better financials 2) The training is more focused on technology and job related skill set, but often ignored the level<br />

of depth in imparting the behavioral and cultural skills. The main contributions of the paper are expatriate training<br />

business drivers in the context of IT global outsourcing, proposing and benchmarking of Expatriate Canonical model<br />

framework, recommendations on delivery models with expatriate management with regard to IT outsourcing<br />

companies.<br />

Keywords: benchmarking, expatriate management, expatriate training, ROI, onsite, offshore, near-shore<br />

1. Introduction<br />

Technological advancement and innovation in the global business landscape is resulting in lot of<br />

challenges and opportunities for the IT industry. Driven by these challenges and opportunities,<br />

organizations are moving across national borders to align the business need to be in competitive market<br />

place to capitalize the wonderful opportunities and establishing foreign-owned subsidiaries, entering into<br />

joint ventures and strategic alliances. In today’s agile market conditions, the C-level executives are<br />

challenged to grab the opportunities, making referenceable customers to sustain in the business with<br />

human talent. To accomplish these goals, the executives are looking for new ways and models to exploit<br />

the current and next generation growth opportunities. National culture has received so much attention<br />

because it is hypothesized to be a constraint on management practice and organization culture (Barry,<br />

2008).To achieve global delivery excellence, organizations are reshaping the traditional business models,<br />

adopting the outsourcing models, deploying and utilizing its global human talent through expatriate<br />

assignments as platform to manage and win the business overseas. International business assignments<br />

are important to the success of the MNCs because they help to build the level of global competence and<br />

global integration within the organization [Paula MaryAnne M and Aparna, 1998). Due to expatriate<br />

assignments by multinational companies, language and mobility parameters have significantly impacted<br />

across the globe. As the world continues to globalize, firms are required to manage an increasingly<br />

diverse workforce with expatriation being just a subset of this challenge.<br />

Global expansion strategies are essential for small, mid and large sized IT organizations to expand and<br />

sustain in the business by recruiting and retaining globally competent people and utilize their managerial<br />

and technical capabilities for a competitive advantage. Expatriate management nowadays is regarded as<br />

one of the tools international organizations use to enhance global integration (Evans, Pucik, and Barsoux,<br />

2002). To lead these international operations successfully, flexible and highly skilled workforce with<br />

global mindset and strategic vision is needed and as such the expatriate Managers are on boarded.<br />

Successful Expatriate assignments drive revenue, value and growth to the organization and have always<br />

been a driving and determining factor for the success of the organization’s overseas operations. These<br />

expatriate managers become the cornerstones on which the international operations are built as they<br />

502


Swathi Duppada and Rama Chandra Aryasri<br />

help in informal control mechanism, for knowledge transfer, and for international team development.<br />

Hence, an emerging view of foreign assignments is that it may create a unique competitive advantage<br />

which is difficult to imitate by competitors.<br />

The research study revealed that the multinational organizations are structured and governed based on<br />

their working model and relationships. There are several types of global assignments that exist based on<br />

the business need and trends, hence the expatriates will be sent to foreign locations with different<br />

duration types – Short term, mid term and long term. Based on the survey analysis, most of the<br />

multinational companies classify the short term assignments duration may include 1-3 months, midterm<br />

assignments duration may include 3-6 months and long term duration may include more than 6 months.<br />

In this paper, the research is focused on outsourcing business model in IT organizations and expatriates<br />

transformation.<br />

2. Literature survey<br />

According to CIPD/KPMG labor market outlook report on skills, migration and off-shoring, almost half<br />

(45%) of the 600 employers surveyed report vacancies that are hard to fill, with 21% saying they are<br />

recruiting migrant workers for engineering vacancies, and 18% for both IT and accountancy/finance<br />

positions. As a result, almost one-fifth (17%) intend to recruit migrant workers in the third quarter of 2010.<br />

In the past three months, one in five (21%) of employers surveyed recruited migrant workers with over a<br />

third (37%) of these workers being recruited from outside the <strong>European</strong> Economic Area (EEA). More than<br />

half of migrant workers hired by the financial sector come from outside the EEA. The findings also point<br />

to greater off shoring activity. Nearly one in ten (9%) private sector companies plan to offshore jobs in the<br />

12 months to June 2011. Of those planning to offshore UK jobs, two-thirds (65%) intend to offshore to<br />

India, a third to China (36%) and three in ten to eastern Europe (29%). The most common functions<br />

outsourced by employers include call centres (55%), IT (51%), and finance (49%) (KPMG & CIPD, 2010).<br />

A fundamental challenge faced by multinational companies today is how to ensure that managers<br />

develop not only an overview of the organization in its entirety, but also a feel for international business<br />

(Gooderham & Nordhaug, 2003). A lot of research still continues to view the globalization topic is a work<br />

in progress research. Only a few firms are considered to have developed an effective capability to locate,<br />

source, manage human resources any where in the world (Lewin and Volberda, 2003), and multi national<br />

enterprises continue to have assets, sales, ownership of work forces and control concentrated in home<br />

countries or regions (Rugman & Verbeke, 2004),). The study of global staffing has traditionally<br />

concentrated on resourcing key positions within Multinational enterprises and top management top key<br />

positions at Head Quarters and subsidiary locations, generally idiosyncratic mixes of strategy rather than<br />

any logical progression of focus related to process of globalization Harzing, 2001, Harzing & Van<br />

Ruysseveldt, 2004) . The literature gives most attention to recruitment and selection of expatriates and<br />

international managers, talent management at HQ or local level, and more flexible forms of international<br />

business travelers, virtual teams and impatriates (Scullion & Collings, 2006). Briscoe and Schuler noted<br />

that the definition of international employee inside organizations continues to expand: “… the tradition of<br />

referring to all international employees as expatriates – or even international assignees – fall short of the<br />

need for international HR practitioners to understand the options available…and fit them to evolving<br />

international business strategies [(Briscoe and Schuler, 2004)”. In the context of international resourcing,<br />

this now includes a fragmentary group of individuals, ranging from: contract expatriates (Baruch and<br />

Altman, 2002), assignees on short term or intermediate term foreign postings (Morley and Hearty, 2004,<br />

Mayerhofer, Hartmann, and Herbert, 2004) ; international transferees (moving from one subsidiary to<br />

another) (Millar and Salt, 2006); virtual international employees active in cross-border project teams<br />

(Janssens and Brett , 2006) ; self-initiated movers who live in a third country but are willing to work for a<br />

multinational (Tharenou, 2003); immigrants actively and passively attracted to a national labor market<br />

and domestically based employees in a service centre but dealing with overseas customers, suppliers<br />

and partners on a regular basis . Millar and Salt draw attention to a number of factors that have increased<br />

demand for new forms of international mobility (Millar and Salt, 2006) : the need for skilled expatriates to<br />

help build new international markets (Findlay, Jowett and Skeldon, 2000) ; temporary and short term<br />

access to specialized talent in sending countries to assist the execution of overseas projects (Minbaeva &<br />

Michailova, 2004 and Hocking, Brown and Harzing , 2004), ; and the need for highly mobile elites of<br />

management to perform boundary-spanning roles to help build social networks and facilitate the<br />

exchange of knowledge (Tushman & Scanlan, 2005).<br />

3. Outsourcing delivery patterns in it organizations<br />

According to David J. Bryce and Michael Useem, outsourcing is defined as an outside company’s<br />

provision of the products or services associated with a major function or activity of a user organization<br />

503


Swathi Duppada and Rama Chandra Aryasri<br />

(David & Michael, 1998) . The definition of outsourcing is not yet standardized; however, the outsourcing<br />

term is much widely used in information technology business. According to Wikipedia, outsourcing is<br />

often viewed as involving the contracting out of a business function - commonly one previously performed<br />

in-house - to an external provider (Overby, 2007). In February 2004, the members of the President<br />

Bush’s council of economic advisers stated: outsourcing of professional services is a prominent example<br />

of a new type of trade” (Gregory, Kristin, Harvey, 2004). The following are the successful outsourcing<br />

implementation patterns due to reduced cost, access to most talented resources at cheater cost, 24 X 7<br />

productivity,<br />

3.1 Onsite delivery model / staff augmentation model<br />

In onsite delivery model, the customer/client would request for skilled expatriates from offshore to work at<br />

their location. In nutshell, the customer work will be done by expatriates at client location with lesser cost.<br />

Normally this model will be adopted when the continuous involvement of client is required at every stage<br />

and security of the data and information is required. Also this is best fitted when the requirements are not<br />

very clear. This model is suitable for short term engagements like strategy consulting, due diligence,<br />

architecture consulting, Roadmap Initiatives and will increase the top line revenues of the offshoring firm<br />

providing services to client.<br />

Figure 1: One view of delivery models<br />

3.2 Offsite delivery model / near shore model<br />

When the client needs more number of expatriates to deliver and if the client cannot extend the facilities<br />

to the expatriates, then the client would prefer the offsite delivery model as the expatriates can use the<br />

foreign/subsidiary office to deliver the projects. The only difference from Onsite delivery models is that the<br />

majority of the expatriates will work from the same country as the client location /city/near city. This model<br />

will enable expatriates to come and conduct face to face discussions and is suitable for medium sized<br />

project.<br />

504


3.3 Offshore delivery model<br />

Swathi Duppada and Rama Chandra Aryasri<br />

When the projects need to be done at cheaper / affordable cost, this model is ideally suitable, however<br />

this model requires clear planning and should possess clear objectives. In this model, the project delivery<br />

is accomplished at offshore (which is in different country). This type delivery model requires frequent<br />

communications with telephone calls and emails as there is no face to face communication.<br />

3.4 Onsite-offshore delivery model<br />

In onsite-offshore delivery model, the work is distributed between Onsite (Client location/host country)<br />

and offshore (home country), hence the client/customer gets most benefitted by taking the advantage of<br />

talent and time from offshore and expatriates from the same offshoring partner will be available at onsite.<br />

The normal distribution ratio of offshore and onsite is 70% and 30%. This model is most successful model<br />

adopted by the Multi National companies to deliver projects better and cheaper. This model is preferred<br />

by the customer as they see significant benefits - 24 X 7 productivity, Most talented offshore resources at<br />

cheaper cost, Access to well trained expatriated to communicate on day to day basis and less burden<br />

with the administration issues.<br />

4. Expatriate transformation canonical framework and benchmarking<br />

The suggested Expatriate transformation canonical framework is very useful to the IT organizations in<br />

transforming the offshore talent for global needs. The purpose of this framework is minimizing the<br />

expatriate dependencies across different business units (verticals) and competency units (technology<br />

units) in different geographic locations. The following diagram depicts the expatriate movement between<br />

locations based on the project needs.<br />

Figure 2: Expatriate transformation canonical framework<br />

Notation: Req - Expatriate Request Res Expatriate Response, MNC – Multi-National Company,<br />

APAC – Asia Pacific, T&M – Time & Material<br />

In order to excel in providing best-in-class expatriates, the canonical framework should be used with the<br />

suitable adapters for the transformation of the expatriate request and expatriate response. Most IT<br />

505


Swathi Duppada and Rama Chandra Aryasri<br />

organizations traditionally follow Hub and Spoke resource model for resources. The business units would<br />

communicate the resource requirements to corporate resource pool through the business unit talent<br />

management team. Hence the first transformation takes place from business unit expatriate resource<br />

request message to pool to transform organization protocol / common organization language. The<br />

request has several input parameters like duration, type, skill, role, etc. These parameters will be used in<br />

determining the right expatriate for the assignment request. This framework has the flexibility to add new<br />

types of expatriate requests.<br />

Figure 3: Expatriate movement beyond boundaries<br />

Typically expatriate management/ training and other related activities will have their own business<br />

process and each of these processes will have process steps. Most likely each process step will deliver<br />

some kind of services. These services may internally use all sorts of formats; hence there is a need to<br />

transform the messages to a common message format.<br />

5. Expatriate process benchmarking, tool and analysis<br />

According to Gartner survey report in December 2010, India is on the top list for offshore services<br />

(Gartner, 2010). To investigate the Expatriate management and training practices for HR excellence to<br />

meet the on demand needs, the authors have selected the MNCs having operations across the globe.<br />

The extensive research on the organizations is conducted to study the patterns and practices of<br />

expatriation. The following are samples of it some of the organizations reveal the need to study, analyze,<br />

benchmark the expatriate practices based on the annual reports.<br />

100.00%<br />

90.00%<br />

80.00%<br />

70.00%<br />

60.00%<br />

50.00%<br />

40.00%<br />

30.00%<br />

20.00%<br />

10.00%<br />

0.00%<br />

Cognizant Revenu Chart for past 4 years<br />

79.10% 79.10%<br />

18.50%<br />

19.20%<br />

82.82%<br />

16.06%<br />

86.19%<br />

12.91%<br />

2.40% 1.70% 1.12% 0.90%<br />

2009 2008 2007 2006<br />

Year<br />

Figure 4: Cognizant revenue contributions (Cognizant, 2009)<br />

506<br />

NA<br />

EUROPE<br />

APAC


% of Revenue<br />

70.00%<br />

60.00%<br />

50.00%<br />

40.00%<br />

30.00%<br />

20.00%<br />

10.00%<br />

0.00%<br />

Swathi Duppada and Rama Chandra Aryasri<br />

65.80%<br />

63.20%<br />

Infosys Revenue chart for past 2 years<br />

26.40%<br />

23.00%<br />

1.20% 1.30%<br />

10.00% 9.10%<br />

North America Europe India ROW<br />

Region<br />

Figure 5: Infosys revenue contributions (Infosys, 2009<br />

60<br />

30<br />

20<br />

10<br />

0<br />

52.8<br />

51.38<br />

50<br />

% of Revenue<br />

40<br />

16.18<br />

TCS % of Revunue for the Y 2009-10 and 2009-08<br />

18.99<br />

10.49 10.53<br />

8.65 7.85<br />

5.24 4.75 4.72 4.71<br />

1.92 1.79<br />

North America UK Europe India APAC Iberoamerica Middle Ease<br />

and Africa<br />

Region<br />

Figure 6: TCS revenue contributions (TCS, 2009) Figure 7: Wipro revenue contributions (Wipro, 2009)<br />

Figure 8: Expatriate process benchmarking<br />

2009-10<br />

2008-09<br />

507<br />

% of Revenue COntribution<br />

70%<br />

60%<br />

50%<br />

40%<br />

30%<br />

20%<br />

10%<br />

0%<br />

2009-10<br />

2008-09<br />

Wipro Revenue chart for past 2 years<br />

2009-10<br />

2008-09<br />

North America Europe Japan ROW<br />

Region and Year


Swathi Duppada and Rama Chandra Aryasri<br />

For these interviews, companies that previously had addressed or involved in addressing the<br />

transformation of their HR functions were selected. A 3-step process is used to conduct the study on the<br />

expatriate processes with multi channel methods. As a first step, the details of the existing Expatriate<br />

Management and training practices and models of the organizations are obtained. The data collection<br />

analysis activity began by identifying the benchmarking partner followed by making available the online<br />

survey weighted questionnaire about their expatriate life cycle process at their organization.<br />

The questionnaire includes – pre-assignment (to capture the drivers of assignment from the perspective<br />

of assignment key stakeholders), on-assignment (to monitor the deliverables and the progress of the<br />

assignment) and post assignment (to help assignment success and synergy to the organization). The<br />

criterion that was set for organizations to be a benchmarking partner is that the organization should have<br />

at least global presence, specifically in North America (NA), Europe and Asia-Pacific which includes<br />

India, in conducting IT business.<br />

Table 1: Normalized scores of the detailed survey<br />

BENCHMARKING PARTNER<br />

Benchmarking Framework Attribute 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17<br />

Structure of the organization 0.87 0.82 0.81 0.86 0.86 0.9 0.83 0.86 0.82 0.89 0.82 0.89 0.84 0.87 0.86 0.82 0.86<br />

Expatriate Request transformation 0.84 0.85 0.79 0.85 0.86 0.9 0.84 0.84 0.84 0.87 0.82 0.84 0.85 0.86 0.85 0.84 0.86<br />

Expat Planning Automation 0.83 0.82 0.79 0.82 0.84 0.87 0.82 0.82 0.82 0.84 0.79 0.83 0.84 0.84 0.84 0.81 0.83<br />

Expat Screening Automation 0.9 0.69 0.67 0.75 0.69 0.72 0.68 0.76 0.68 0.74 0.68 0.85 0.69 0.79 0.69 0.68 0.77<br />

Expat Interviewing and Automation 0.87 0.73 0.71 0.76 0.74 0.69 0.74 0.76 0.73 0.73 0.73 0.87 0.72 0.74 0.74 0.73 0.71<br />

Expatriate Selection Automation 0.81 0.87 0.84 0.86 0.89 0.91 0.87 0.84 0.87 0.92 0.85 0.81 0.89 0.89 0.83 0.87 0.85<br />

Immigration Process Automation 0.74 0.84 0.82 0.91 0.85 0.89 0.84 0.88 0.83 0.89 0.83 0.79 0.85 0.88 0.85 0.83 0.81<br />

Training Automation 0.89 0.82 0.82 0.81 0.87 0.93 0.83 0.85 0.82 0.91 0.82 0.85 0.84 0.87 0.84 0.82 0.84<br />

Depature/Deputation/Onsite transfer automation 0.81 0.81 0.81 0.79 0.81 0.88 0.83 0.83 0.81 0.85 0.81 0.81 0.81 0.86 0.81 0.81 0.79<br />

Arrival Support Automation 0.83 0.78 0.76 0.82 0.79 0.92 0.79 0.85 0.76 0.91 0.76 0.83 0.79 0.89 0.79 0.76 0.75<br />

Strategic Partnering 0.82 0.75 0.69 0.78 0.75 0.95 0.75 0.79 0.74 0.82 0.71 0.82 0.74 0.81 0.74 0.71 0.76<br />

Expatriate Service Development 0.78 0.68 0.52 0.75 0.79 0.7 0.69 0.78 0.67 0.81 0.62 0.74 0.71 0.79 0.71 0.62 0.77<br />

Sum of the Normalized Scores 9.99 9.46 9.03 9.76 9.74 10.26 9.51 9.86 9.39 10.18 9.24 9.93 9.57 10.09 9.55 9.3 9.6<br />

Score Ranking 4 13 17 7 8 1 12 6 14 2 16 5 10 3 11 15 9<br />

The above graphs clearly show that the organizations with good recruitment practices stand in top<br />

position; thereby the authors have arrived at the best practices in the next section.<br />

In order to find out the results and observations, the original sample included 100 associates from each<br />

firm were identified. With 69% of participants, the data is collected on technology, managerial and<br />

behavioral skills<br />

Normalized Score<br />

10.4<br />

10.2<br />

10<br />

9.8<br />

9.6<br />

9.4<br />

9.2<br />

9<br />

8.8<br />

8.6<br />

8.4<br />

Sum of the Normalized Scores V/S Benchmarking Partner<br />

B1 B2 B3 B4 B5 B6 B7 B8 B9 B10 B11 B12 B13 B14 B15 B16 B17<br />

Benchmarking Partner<br />

Sum of the Normalized Scores<br />

Normalized Values<br />

18<br />

16<br />

14<br />

12<br />

10<br />

8<br />

6<br />

4<br />

2<br />

0<br />

BP Partner Ranking with Normalized Values<br />

Sum of the Normalized Scores Score Ranking<br />

B1 B2 B3 B4 B5 B6 B7 B8 B9 B10 B11 B12 B13 B14 B15 B16 B17<br />

Benchmarking Partner<br />

Figure 9: Normalized scores with survey results Figure 10: Ranking with survey results<br />

Statistical analysis are performed to obtain the top most recruitment metrics – Full productivity, Quality of<br />

Service, Quality of Promotion and Knowledge Management of expatriates for mid to long term duration.<br />

The results are captured after the research, analysis and screening using Expatriate transformation<br />

framework.<br />

Quality of promotion is an important metric for branding. This parameter will have increased influence of<br />

people, processes and opportunities of the expatriate representing organization.<br />

6. Future trends<br />

In the days to come expatriate processes would be completely modernized and automated. The use of<br />

technology in the sector continues to increase at a remarkably faster pace with technologies hence;<br />

organizations have to be on top of technology and competencies along with the deep domain knowledge<br />

to address the future expatriate resource fulfillment needs.<br />

508


Full Productivity<br />

8<br />

7<br />

6<br />

5<br />

4<br />

3<br />

2<br />

1<br />

0<br />

-1<br />

-2<br />

-3<br />

Average Full Productivity w ith/out ETCFABM<br />

Swathi Duppada and Rama Chandra Aryasri<br />

With ETCFABM Adoption Without ETCFABM Adoption<br />

Intial Offshore EBMF Adoption Onsite Log term Onsite<br />

Time<br />

Quality of Service<br />

9<br />

8<br />

7<br />

6<br />

5<br />

4<br />

3<br />

2<br />

1<br />

0<br />

Average Quality of Service w ith/out ETCFABM<br />

With ETCFABM Adoption Without ETCFABM Adoption<br />

Intial Offshore EBMF<br />

Adoption<br />

Onsite Log term<br />

Onsite<br />

Time<br />

8 - Superior 6- Meritorious 4-Acheivable 2- Reasonable 0-Nornal<br />

Figure 11: Average full productivity Figure 12: Average quality of service<br />

Quality of Promotion<br />

4.5<br />

4<br />

3.5<br />

3<br />

2.5<br />

2<br />

1.5<br />

1<br />

0.5<br />

Quality of Promotion with/out ETCFABM<br />

With ETCFABM Adoption Without ETCFABM Adoption<br />

0<br />

Intial Offshore<br />

Time<br />

EBMF Adoption Onsite<br />

5 - Excellent 4- Very Good 3-Good 2-Reasonale 1-Nornal<br />

Knowledge Management<br />

6<br />

5<br />

4<br />

3<br />

2<br />

1<br />

0<br />

Knowledge Management with/out ETCFABM<br />

With ETCFABM Adoption Without ETCFABM Adoption<br />

Intial Offshore EBMF Adoption Onsite<br />

Time / Locale<br />

5 - Excellent 4- Very Good 3-Good 2-Reasonale 1-Nornal<br />

Figure 13: Average quality of promotion Figure 14: Average knowledge management<br />

Expatriate processes would become more efficient and robust and fully integrated with the overall<br />

organizational objectives. Firms will have to pay high attention to ensure that the expatriate solutions<br />

address the employer branding and regulatory compliances issues. The focus on building passive<br />

candidate pool and diverse recruitment would increase tremendously. High attention will be paid towards<br />

achieving collaboration between the hiring Manager and the recruiter.<br />

References<br />

Barry Gerhart, (2008) Cross Cultural management Research: Assumptions, Evidence and Suggested Directions,<br />

International Journal of Cross Cultural Management.<br />

Baruch, Y. and Altman, Y. (2002) Expatriation and Repatriation in MNCs : a Taxonomy, Journal Human Resource<br />

Management, Pg 239-259<br />

Briscoe, D. and Schuler, R.S. (2004) International Human Resource Management”, Second Edition, New York:<br />

Routledge.<br />

Cognizant Report. (2009) Cognizant Annual Report, http://investors.cognizant.com/index.php?s=52<br />

David J.Bryce and Michael Useem.(1998) The impact of corporate outsourcing on company value, <strong>European</strong><br />

Management Journal, Vol. 16, No.6, pp.635-643.<br />

Evans, P, Pucik, V and Barsoux, J.L, (2002) The Global Challenge: International Human Resource Management,<br />

NewYork: McGraw-Hill/Irwin.<br />

Findlay, A.M., Li, F.L.N., Jowett, A.J. and Skeldon, R. (2000) Skilled international migration and the global city: a<br />

study of expatriates in Hong Kong”. Applied Geography, Pg: 277- 304.<br />

Gartner Report. (2010) Gartner's 30 Leading Locations for Offshore Services, 2010-2011,<br />

http://www.gartner.com/it/page.jsp?id=1502714.<br />

Gooderham, P.N and Nordhaug, O (2003) International Management: Cross-boundary challenge, Oxford, Blackwell<br />

Publishing.<br />

Harzing, A.W.K (2001) Who’s in charge? An empirical study of executive staffing practices in Foreign Multinationals,<br />

Journal of Human Resource Management, 40 139-58.<br />

509


Swathi Duppada and Rama Chandra Aryasri<br />

Harzing, A.W.K and J.Van Ruysseveldt. (2004) Composing and International Staff, International Human Resource<br />

Management, Second Ed.<br />

Hocking, J.B., Brown, M. and Harzing, A.-W. (2004),A knowledge transfer perspective of strategic assignment<br />

purposes and their path-dependent outcomes, International Journal of Human Resource Management, 15 (3):<br />

565-586.<br />

Infosys Report. (2009) Infosys Annual Report, http://www.infosys.com/investors/pages/index.aspx<br />

Janssens, M. and Brett, J.M. (2006), Cultural intelligence in global teams: a fusion model of collaboration, Group and<br />

Organization Management, 31 (1): 124-153.<br />

KPMG & CIPD Quarterly Survey Report, (2010)Labor Market Outlook,<br />

http://www.cipd.co.uk/subjects/hrpract/hrtrends/_qtrends.htm.<br />

Lewin, A.Y. and Volberda, H.W (2003) Beyond Adaptation – Selection Research: Organizing Seld-Renewal in Co-<br />

Mayerhofer, H., Hartmann, L.C. and Herbert, A. (2004) Career management issues for flexpatriate international<br />

staff”, Thunderbird International Business Review, Pg: 647- 666.<br />

Millar, J. and Salt, J. (2006) The mobility of expertise in transnational corporations, Migration Research Unit. Working<br />

Paper. University College London.<br />

Minbaeva, D.B. and Michailova, S. (2004) Knowledge transfer and expatriation in multinational corporations: the role<br />

of disseminative capacity, Employee Relations, 26 (6): 663-679<br />

Morley, M. and Hearty, N. (2004) International Assignment and global careers, Thunderbird International Business<br />

Review, Pg: 633-646.<br />

N. Gregory Mankiw, Kristin J. Forbes, Harvey S. Rosen. (2004) The Economic Report of the President”, Washington,<br />

D.C.<br />

Overby, S. (2007) ABC: An Introduction to Outsourcing, www. CIO.com.<br />

Paula M.Caligiuri, MaryAnne M and Aparna Joshi (1998) Families on Global Assignments: Applying work/family<br />

theories abroad, Current Topics in Management, Volume 3, pages 313-328.<br />

Rugman, A. and Verbeke, A (2004) A perspective on regional and global strategies of multinational enterprises,<br />

Journal of International Business Studies, 35: 3-18.<br />

Scullion, H and Collings, D.G. (2006), “Global Stafing”, London: Routledge.<br />

TCS Report. (2009) TCS Annual Report,<br />

http://www.tcs.com/investors/Documents/Annual%20Reports/TCS_Annual_Report_2009-2010.pdf<br />

Tharenou, P. (2003) The initial development of receptivity to working abroad: self-initiated international work<br />

opportunities in young graduate employees, Journal of Occupational and Organizational Psychology, 76: 489-<br />

515.<br />

Tushman, M.I. and Scanlan, T.J. (2005) Boundary spanning individuals: their role in information transfer and their<br />

antecedents , Academy of Management Journal, 24 (2): 289-305.<br />

Wipro Report. (2009) Wipro Annual Report http://www.wipro.com/corporate/investors/pdf-files/annualreport/wipro_annual_report.pdf.<br />

510


The Diffusion of Mobile Phones for Business and Information<br />

Management in Kenya<br />

Wakari Gikenye<br />

University of Zululand, South Africa<br />

wagikenye@yahoo.com<br />

Abstract: Diffusion of innovation is the process by which an innovation is adopted and gains acceptance by<br />

members of a certain community, innovations that are perceived by individuals as having greater relative advantage,<br />

compatibility and less complexity may be adopted more rapidly than others (Rogers 2003:5 :16). Information and<br />

communication technologies (ICTs) like the computer and the internet have been widely adopted by people in<br />

western countries, while in developing countries they have taken much longer to be adopted. Mobile phones ICT on<br />

the other hand have been widely adopted by both the western and developing countries in the last decade. In Kenya,<br />

for instance, the number of subscribers had reached 20 million by March 2010, and still growing (Communication<br />

commission of Kenya 2010). The rapid spread and use of mobile phone technology has in the last three years been<br />

extended to the use of innovative mobile money transfer services both of which have increased the speed of<br />

economic activities for Kenyans, including the low income people who were not served by financial institutions like<br />

banks and the post office network. Money transfer services were first introduced in Kenya in 2007 in the form of ‘Mpesa’,<br />

a financial transaction medium, which has had phenomenal growth since its introduction and has greatly<br />

impacted economic activities. The M-pesa mobile money transfer service has been described as a revolutionary way<br />

of making cell phone technology work to serve a basic need by providing a fast, safe, cheaper and easy way to<br />

transfer money (Vodacom 2010). This paper examines the rapid diffusion of mobile phones and related mobile<br />

money transfer services in Kenya. It is based on an on-going research study on “The diffusion of ICTs in the informal<br />

sector in Kenya”.<br />

Keywords: diffusion of ICTs, mobile phones adoption, money transfer services, Kenya<br />

1. Introduction and background<br />

Diffusion of innovation is the process in which an innovation is adopted and gains acceptance by<br />

members of a certain community, innovations that are perceived by individuals as having greater relative<br />

advantage, compatibility and less complexity may be adopted more rapidly than others ((Rogers<br />

2003:5:16). Information and communication technologies (ICTs) like the computer and the internet have<br />

been widely adopted by people in the western countries, while in developing countries they are taking<br />

much longer to be adopted. Mobile phones ICT on the other hand have been widely adopted by both the<br />

western and developing countries during the last ten years. In Kenya, by March 2010 mobile phone<br />

subscription had reached 20 million and still growing (Communication commission of Kenya 2010). The<br />

rapid spread of mobile phone technology in Kenya has been extended to the use of innovative money<br />

transfer services. This has impacted the speed at which economic activities are carried out especially for<br />

those people who were not served by financial institutions like the banks.<br />

Money transfer services were first introduced in Kenya in 2007 in the form of ‘M-pesa’, it is a non-bank<br />

service where customers exchange cash at M-pesa agents in return for an e-money account, the virtual<br />

account can then be used for making payments, storing funds for future use and transferring funds to<br />

other phone users (Standard Chartered 2009/2010). The M-pesa money transfer service has been<br />

described as a revolutionary way of making cell phone technology work to serve a basic need by<br />

providing a fast, safe, cheaper and easy way to transfer money (Vodacom 2010). It was initially started as<br />

a mobile money transfer service to improve the efficiency of microfinance by using mobile technology to<br />

make transactions faster, cheaper and more secure (Geach, 2007:4) Other mobile money transfer<br />

services introduced later are Zap by Zain which is Safaricom’s main competitor in Kenya, ‘Yu Cash’ for<br />

YU mobile phone operator and orange money from Orange mobile phone operator, these services are<br />

heavily used for financial transactions and have also created jobs for agents in the many outlets across<br />

the country. As Kalba (2008:632) has rightly observed, mobile phones are “… spreading ubiquitously<br />

across the planet and are the latest phase of globalization’’ having out-diffused virtually every prior<br />

technology, whether TV sets, radios, wrist watches, fixed lines computers, Internet etc.<br />

This paper examines the rapid diffusion of mobile phones and mobile money transfer services and their<br />

effect on the information and business management in Kenya. It is based on an on-going research study<br />

“The diffusion of ICTs in the informal sector in Kenya”, the purpose of the study is establish the status of<br />

ICT penetration (or lack thereof) in the micro and small enterprise (MSE) sector in Kenya, the terms<br />

informal sector and MSEs sector are used interchangeably in the study. The study attempts to answer<br />

511


Wakari Gikenye<br />

the following research questions: What is the status of ICT penetration/access, use and impact (or lack<br />

thereof) in the MSE sector in Kenya?, What is the rate and level of ICT diffusion in the informal sector in<br />

Kenya?, What types of ICTs are being used by the MSE traders in Kenya, and what impact if any do they<br />

have on the businesses?, What challenges if any hinder the awareness and use of ICTs in the informal<br />

sector in Kenya?, What is the government involvement in putting the required infrastructure for ICT use in<br />

the informal sector Kenya.<br />

2. Review of relevant literature<br />

The rapid diffusion of the mobile phone in the last ten years can be explained by; among other things, the<br />

drop in the price of mobile handsets to within reach of those with low incomes, the drop in mobile tariffs<br />

as a result of stiff competition between mobile phone operators, mobile phones also require only basic<br />

literacy to use and this makes them accessible to a larger proportion of the population, they are easy to<br />

use and adaptable and for those without electricity phone–charging kiosks have quickly come up in small<br />

towns and shopping centers while solar-powered phones have also been introduced into the market<br />

(CCK 2010). The low cost of prepaid calling cards has also contributed to the rapid diffusion (Standard<br />

Chartered 2009/2010:61; Communication Commission of Kenya, 2010).<br />

Mobile ICT technology has made it possible to extend financial transfer services to millions of people<br />

including the low income at relatively low cost (Kimenyi and Ndungu 2009:2). Mas and Radcliffe (2010:2)<br />

have observed that, there nearly five times the number of M-pesa outlets than the total number of all the<br />

other financial services i.e. Post Bank branches, post offices, bank branches and automated teller<br />

machines all put together in the country. This is because unlike conventional banking branches that are<br />

only located in major urban centers due to high running expenses, M-pesa agents are the ones who<br />

decide on the most strategic points to locate their service outlets since they are easy to put up and to<br />

maintain, this has dramatically reduced the cost of delivering financial services, with the benefits being<br />

transferred to the users (Njenga nd :5).<br />

Mpesa has shown impressive growth in the utilization of mobile payments within a short period of time,<br />

with the number of monthly transactions having increased from 354,298 in July 2007 four months after it<br />

was introduced to 16,747,419 in July 2009, an increase of 4627 percent, the value of monthly<br />

transactions also rose from Kenya shillings 1.065 billion (Us $ 14.2 million) in July 2007 to 40.176 billion<br />

(Us $535.6 million) in July 2009, a growth of 3671 percent, these financial transactions generated<br />

substantial returns, and created job opportunities with about 12,300 agents in the country (Kimenyi and<br />

Ndungu 2009:3).<br />

Factors that have been responsible for the highly successful money transfer services in Kenya are; the<br />

fast adoption of mobile phones, the unmet need to access financial services as well as a strong demand<br />

for domestic remittances, such needs used to be met through informal and risky methods like sending<br />

money through relatives and friends, or matatus (shared taxi), buses and couriers by those working in the<br />

urban areas to their relatives in the rural areas, (Omwansa 2009).<br />

M-pesa therefore may have been introduced at the right time, to meet the needs of accessing financial<br />

services especially by those who were not served by the formal banking institutions. The cost of mobile<br />

money service is relatively low compared to the earlier methods, it is also faster more convenient and<br />

safer than cash. The adoption of mobile-based money transfer service has speeded up access to<br />

financial services not just for the unbanked but also for the banked as a preferred avenue of transactions<br />

(Daily Nation September 29 2010). The biggest beneficiaries, however, are the MSE traders who have<br />

discovered a new time-saving service leaving them with more time to work on their businesses<br />

(Omwansa 2009:111). The mobile money transfer have services also offered protection and safety by<br />

enabling instant and on-demand payments (Mas and Radcliffe 2010:5). Other innovative successful<br />

money transfer services in Africa are Wizzit in South Africa and Glofirst card in Nigeria which are used to<br />

draw cash, transfer money, make purchases, and check bank statements using the mobile phone<br />

(Standard Chartered 2009/2010:61).<br />

2.1 Methodology: How do MSEs in Kenya get information for their business<br />

management?<br />

According to Rogers’ theory of diffusion (2003:1), getting a new innovation adopted even when it has<br />

obvious advantages is difficult and many innovations require a period of many years to gain wide<br />

adoption. The slow adoption of computer and the internet for information provision in developing<br />

512


Wakari Gikenye<br />

countries provide examples while adoption and use for the same has been much faster in Western<br />

countries. Technology acceptance model (TAM) developed by Davis (1989) stresses perceived<br />

usefulness (PU) and perceived ease of use (PEOU) of an innovation (Khasawney and Ibrahim 2008:137).<br />

A combination of social, economic and psychological factors might explain how different communities<br />

become aware of and adopt particular innovations while they take much longer to perceive the<br />

usefulness and ease of use to others.<br />

A survey was carried out in two provinces in Kenya where a total of 377 MSEs were selected using a<br />

combination of multistage sampling technique consisting of purposive, cluster and random sampling. The<br />

two provinces selected are Nairobi and Central provinces representing an urban and a rural area<br />

respectively. The MSEs consisted of retail shops (mainly dealing with clothing and footwear), phone<br />

shops, auto-spare shops, furniture and hardware shops, and curio and horticultural traders. The clothing<br />

and footwear retail shops, phone shops and electronic shops were more concentrated in the city center<br />

while hardware, furniture, hair salons and construction material stores were found in the market centers.<br />

Curio traders were found in the up-market Village Market in Nairobi province while horticultural traders<br />

were based next to Jomo Kenyatta International Airport.<br />

Data was collected using comprehensive structured and non-structured questionnaires, to gather<br />

information from MSE owners, managers and employees on their use of ICTs for business management.<br />

The questions were based on their use, attitudes and impact of ICTs. Due to the informal nature of the<br />

business environment in the MSE sector, researcher-administered questionnaires were used.<br />

3. Results<br />

The results are presented in sections 3.1. to 3.7<br />

3.1 Gender composition, educational level and age of respondents<br />

The sampled MSE traders were 49 percent men and 51 percent women, making the gender balance<br />

quite close. As shown in table 1 the educational level of the traders varied from primary to degree level,<br />

with more than 27% (104 respondents) of the respondents having gone up to degree/diploma level and<br />

48% (182) up to secondary school level. 16% (60 respondents) had undergone various types of post<br />

secondary training, while over 6% (25 respondents) were primary school leavers. Majority of respondents<br />

had attained secondary school education and above.<br />

Table 1: Educational level of respondents<br />

Frequency Percent<br />

Primary Level 25 6.6<br />

Secondary Level 182 48.3<br />

Post Secondary Training 60 15.9<br />

Diploma/degree/Masters 104 27.6<br />

No answer 6 1.6<br />

377 100<br />

Table 2 shows the majority of the respondents are relatively young people between the age of 20 to 45,<br />

with 75 percent being under 35 years, these are young people trying to earn a living from MSE trade,<br />

most of whom are school leavers who failed to get jobs in the formal sector.<br />

Table 2: Age of respondents<br />

Frequency Percent<br />

Below 25 years 100 27<br />

25-35 years 180 48<br />

36-45 years 93 25<br />

Over 45 years 1 .3<br />

No answer 3 .8<br />

Total 377 377<br />

513


3.2 Size of MSEs in the study<br />

Wakari Gikenye<br />

The MSEs in the study can be divided into three categories based on their size and the premises they<br />

occupy. The first category operates from permanent buildings and enjoys infrastructural facilities such as<br />

water, electricity, they also pay relatively higher rents. They enjoy relative permanency and their capital<br />

outlay is also relatively higher. Examples of these are horticultural exporters and electronic and autospare<br />

shops. Most MSEs in this category use ICTs such as the computer, Internet and email for their<br />

businesses.<br />

The second category of MSEs operates micro-sized businesses in shared premises. Despite being<br />

housed in permanent premises and sheltered from the vagaries of the weather their businesses and<br />

operations are very small, these are the micro retail stalls found in the main streets of Nairobi.<br />

The third category operates from temporary sheds or stalls or in the open (this is the origin of the term<br />

“Jua Kali’’, Kiswahili for ‘’in the hot sun’’ used generally to refer to informal sector enterprises in Kenya).<br />

Their premises generally lack infrastructural facilities such as running water and electricity. They do not<br />

own or use computers and also lack awareness of ICT developments, examples are curio traders,<br />

grocery stores, food kiosk operators etc.<br />

Over seventy per cent (270) of the MSEs studied had between one and five employees, with 35 percent<br />

(133 respondents) having only one employee, only 7 percent had more than five employees.<br />

3.3 Diffusion and use of ICTs among the MSEs<br />

With the exception of mobile phones, the majority of MSE traders do not use ICTs such as computers for<br />

their businesses. The majority carry on as if such technological developments have nothing to do with<br />

them.<br />

As table three shows twenty six percent (96 respondents) owned a computer, while over 94% (356<br />

respondents) (shown in table 4) reported owning mobile phones. There was also a relatively high number<br />

of respondents with computer skills; 42 percent (158) but who were not using them because they did not<br />

think their businesses need the use of computers and said they could not afford them.<br />

Table 3: Ownership of computers in the MSEs<br />

Frequency Percent<br />

Yes 96 25.5<br />

No 281 74.5<br />

Total 377 100.0<br />

3.4 Diffusion and use of mobile phones and money transfer services for business<br />

management<br />

Findings shown in table 4 show that the mobile phone is quite popular among the MSE traders, with over<br />

94% reporting that they owned and used a mobile phone to conduct their businesses. 95 percent (357<br />

respondents) also reported that they were registered M-Pesa users, M-pesa was the pioneer for money<br />

transfer services and has been synonymous with mobile money transfer services in Kenya.<br />

Table 4: Ownership of mobile phone by MSE owners<br />

Frequency Percent<br />

Yes 356 94.4<br />

No 21 5.6<br />

Total 377 100<br />

As shown in table 5, 28 percent (108 respondents) reported using M-Pesa daily, 29% (111 respondents)<br />

use it once or twice a week; 21 % (80 respondents) said they used it whenever the need arose.<br />

Sixty three percent (237 respondents) of the M-pesa users said they used it to send and receive money<br />

from relatives while 79% (296 respondents) said they used it for business transactions. For most<br />

respondents, the mobile phone was their first contact with a phone, having never used the landline. Given<br />

that even the low prices of handsets and cost of airtime is a significant proportion of their earnings, the<br />

514


Wakari Gikenye<br />

high percentage of ownership and use of mobile phones means they value them highly and they would<br />

rather forgo other things to make sure they maintain the use of the mobile phone, which they reported<br />

using heavily for their work. In addition to being useful in business activities mobile phones are also seen<br />

as a status symbol.<br />

Table 5: Frequency on the use of M-pesa (mobile money transfer service)<br />

Frequency Percent<br />

Everyday 108 28<br />

Twice a week 20 5<br />

Once a week 91 24<br />

Once a month 17 5<br />

Other 36 10<br />

When need arises 80 21<br />

No answer 25 7<br />

Total 377 100<br />

Almost 60 per percent (222 respondents) reported that they used the mobile phone for their businesses<br />

transactions, 73% (275 respondents) said they used it for ordering goods, 79% (299 respondents) for<br />

contacting customers, and 50.4% (190) said they had used it to improve their businesses. 72 per cent<br />

(271) said they used the mobile phone for social communication, e.g. contacting friends and relatives.<br />

3.5 What difference has the mobile phone made for MSE traders in Kenya<br />

Fifty one percent (192 respondents) said the use of mobile phones made it easier and faster to<br />

communicate and carry out business transactions at any time as well as communicating with people in far<br />

off places. Fifteen percent (57 respondents) said the mobile phone helped them to get goods delivered<br />

more easily, 31 percent (115 respondents) said the use of mobile phones made it easier to contact and<br />

bring more customers leading to increased revenue.<br />

A small number of respondents (2%; 6 respondents) said that they were what they were because of the<br />

use of mobile phones, i.e. they had been able to start a business and maintain it due to the availability<br />

and use of the mobile phone. They reported that, by using the mobile phone, they were able to know<br />

which market to go to after comparing prices in different markets, before they left home. They were also<br />

able to change plans if the information received on the prices of a particular commodity would not bring<br />

them profits.<br />

Most respondents reported doing businesses they had been doing more easily and conveniently with the<br />

use of the mobile phone, 16%(62 respondents) said they were impressed by the genuineness, speed,<br />

convenience and reliability that the mobile phone had brought and which had resulted in making their<br />

work much easier and convenient, 4 percent (16) went as far as saying they could not do without it as it<br />

had ‘’become everything in their business’’.<br />

Only 2 per cent (6) said they relied on the money transfer service when a customer did not have money<br />

and could send it later, most traders said they prefer face-to-face transactions rather than encouraging<br />

credit services. 4 percent (14), said that M-Pesa is good for security and for carrying the money around<br />

instead of cash, because even if the phone is stolen the money is still safe. 3 per cent (10) said it is good<br />

for emergencies including medical ones due to its speed and convenience. 19 percent (72) said the<br />

money transfer services are conveniently available for more hours unlike banks, and that it is easier to<br />

save small amounts due to instant access, and one need not queue at the bank.<br />

Other uses the mobile phone is used for by the MSE traders are; telling the time (73 percent; 276<br />

respondents) thus replacing the wrist watch, 47 percent (178) use it as a camera, 46 percent (173) used<br />

it to access the Internet, 49 percent (185) use it as a radio.<br />

3.5.1 Problems of using mobile phones by MSE traders in Kenya<br />

Thirty seven percent (140 respondents) said they had problems buying credit for their phones but<br />

nonetheless indicated that it was a necessary cost for doing business, 54 percent per cent (207)<br />

complained about network congestion, a common problem experienced by Safaricom customers the<br />

main telephone service provider in Kenya and M-pesa money transfer service users. The mobile phone<br />

was also reported (46 per cent) to be an easy target by thieves and pickpockets. Only 1 percent (4<br />

515


Wakari Gikenye<br />

respondents) reported having problems with charging their phones, this is significant given the fact that<br />

there are many people in Kenya without access to electricity in their homes.<br />

3.5.2 Information sources for MSE business management<br />

Apart from getting information from mobile phone contacts and fellow traders, 34% (128 respondents) of<br />

MSE traders also reported getting their business information through acquaintances, friends and<br />

relatives, through word of mouth while only 2% reported getting information through government officers.<br />

23 percent said they got information through the mass media while 15% got their information through the<br />

internet.<br />

3.6 Challenges that hinder the awareness and use of ICTs by MSEs in Kenya<br />

As the findings show in 3.2 above the majority of MSEs are small in size with the majority having less<br />

than five employees, most of them are survival outfits, and their size and earnings do not enable the<br />

ownership of ICTs like the computer. As shown in table 6, they are started with minimum preparation as<br />

shown by the amount of starting capital, making the acquisition of expensive ICTs a remote possibility.<br />

Forty eight percent reported starting their businesses with less than 1000,000 Kenya shillings ($1250)<br />

with some having been started with as little as less than 10,000 Kenya shillings ($125).<br />

Table 6: Capital outlay at the beginning of business<br />

Frequency Percent<br />

Less than Kshs 10,000 7 2<br />

Kshs 10,000-20,000 49 13<br />

Kshs 20,000-40,000 41 11<br />

Kshs 40,000-60,000 41 11<br />

Kshs 60,000-100,000 40 11<br />

Over Kshs 100,000 51 14<br />

Do not know 55 15<br />

No answer 93 25<br />

Total 377 100<br />

3.7 Government Involvement in putting the required infrastructure for ICT use<br />

In spite of the efforts the government of Kenya has made to bring the internet closer to Kenyans by<br />

enabling three under sea cables to improve internet connectivity and the reduction on taxes on<br />

computers, the MSE traders do not think the government has done much to assist them in getting the<br />

necessary information for their businesses, they reported that they could not afford computers for their<br />

businesses even for those who have acquired computer skills (3.3 above). Asked why they did not<br />

recognize or appreciate the government efforts to help them get more information through ICTs they said<br />

that that their first priority was to get affordable and more permanent premises where they would<br />

establish their businesses, that the current status of MSE business premises in Kenya is unpredictable<br />

with the City Council’s occasional raids on semi-permanent premises. The MSE traders also felt the<br />

government was too keen to collect taxes and license fees but less eager to give services corresponding<br />

services to enable them establish more productive businesses.<br />

4. Discussion<br />

This study has found that, gender composition for MSE traders studied is the same for men and women,<br />

this is unlike other studies’ findings that there are more women than men involved in MSEs (Okello-Obura<br />

et al 2007:13).<br />

The small size of the MSEs has not endeared them to the use of ICTs like the computer, rather the<br />

majority of the MSEs traders have adopted mobile phones and mobile money transfer services for<br />

management of their businesses, the mobile phone technology has also been well adapted to the local<br />

economic environment for use by both the unbanked and the banked Kenyans.<br />

Due to the convenience resulting from the use of mobile phones and money transfer services and saved<br />

time and money in saved journeys respondents reported that business had become much easier and<br />

faster to run, and was more profitable. Thirty five per cent (131 respondents) said mobile money transfer<br />

services simplified business and social financial transactions, saves time and money, by cutting costs of<br />

516


Wakari Gikenye<br />

traveling and increases the efficiency of doing business. Money transfer services are more flexible and<br />

can be used for saving and withdrawing small amounts of money, they operate for more hours, unlike the<br />

banks, they are cheaper than bank accounts and require no minimum balance.<br />

The mobile phone was reported to have changed business practices for the better as it made it easier to<br />

order, supply and make payments for goods and services using the mobile phone. As summed up in a<br />

Standard Chartered bank report (2009/2010:60) “the mobile phone is having a dramatic effect …and is<br />

proving to be a life-transforming device”, mobile phones have become a key economic driver and are<br />

used in many different ways like marketing of goods and obtaining information, they have reduced travel<br />

time to distant bank branches and have enabled access to banking. According to Kimenyi and Ndungu<br />

(2009:2) Kenya’s mobile phone transfer service is a good example of how low cost approaches using<br />

modern technology can effectively expand the financial services frontier.<br />

Mas and Morawczynski (2009:89) are optimistic that there is more potential use of mobile transfer<br />

services that can be extended to payment of salaries and wages, payment of bills, school fees etc, and<br />

that, although not originally conceived to work as a saving service it has been used as such by the<br />

unbanked even with no interest. Emerging partnerships of various banks with mobile phone providers<br />

taking place in Kenya will make the realization of mobile banking a reality, as well as collectively bring<br />

together the aggregate savings from the wide network of money transfer services’ customers.<br />

While achievements of the mobile phone in the MSE sector are from private sector efforts the<br />

government has played a significant part in providing an enabling environment. This has not been felt in<br />

the adoption and acquisition of other ICTs like the computer and the internet, and the challenge remains<br />

for the government to do more so that the MSE traders can realize the benefits of the other ICTs.<br />

5. Conclusion<br />

This paper has found that, the status of ICT adoption especially that of computers is still low and out of<br />

reach of the MSE traders whose operations continue to remain small. The affordable and easy-to-use<br />

mobile phone technology however has been quickly and extensively adopted and is being used to<br />

provide easier communication of needed information by MSEs in Kenya. This has made businesses more<br />

efficient, less costly and more profitable in the MSE sector. The mobile phone technology has also been<br />

innovatively used to offer financial transfer services whose fast adoption is proof that it has been found<br />

useful and handy not just for the unbanked but also as a service to supplement the banked.<br />

The rapid diffusion of the mobile phone is not only enabling faster and more efficient communication but<br />

is also empowering people including the low income with information and services that were not available<br />

to them before. This will have a positive impact on the economy and with the telephone companies<br />

increasing their activities in the market and the growing partnerships with the banks the Kenyan economy<br />

is bound to grow as a result.<br />

The study also found out that for ICT adoption to be a reality for the MSE traders and indeed for all the<br />

other Kenyans, there is need for concerted efforts by the government, the internet service providers and<br />

for the people themselves to update themselves with what is happening around them and the rest of the<br />

world if they are to catch up and keep abreast with the rest of the world. A lot remains to be done to<br />

enable the realization of ICT benefits so as to maximize information acquisition for business<br />

management. While the mobile phone technology has been innovatively used to provide the MSE traders<br />

and other Kenyans with easier communication and money transfer services it has limitations on the<br />

impact that it can make on the economy of the country even at its best. The government needs to<br />

facilitate the creation of more jobs so as to have a vibrant economy which can increase the demand for<br />

goods and services, including those being produced by MSEs. Lack of jobs is the genesis for the flocking<br />

of the unemployed into the informal sector, where the participants try to eke out a living. Even with the<br />

government efforts to make internet connectivity easier to access and the zero-rating of taxes on<br />

computers and accessories people need to satisfy the basics before they can acquire ICTs.<br />

Appendix 1<br />

Questionnaire<br />

Section A: Personal Information<br />

1. .Name: ________________________________________ (Optional)<br />

517


Wakari Gikenye<br />

2. Age: (1) Below 25 years (2) 25-35 years (3) 36-45 years (3) Over 45<br />

3. Sex: (1) Male (2) Female<br />

4. Marital Status: (1) Single (2) Married (3) Other<br />

5. Educational Level: (1) Primary Education Level (2) Secondary Education Level (3)<br />

Post Secondary Training (specify)_________________(4) Diploma/degree Holder.<br />

Section B: Information about the business enterprise<br />

6. What is the name of your business ________________________________________<br />

7. Are you the (1) Owner (2) Part Owner (3) Manager (4) Employee<br />

8. What are the goals of your business enterprise? _______________________________<br />

9. How many permanent employees does the business have? (1) One only (2) 2-5<br />

(3) 6-10 (4) Over 10 specify how many____________________________________<br />

10. Do you have employees that are not permanent? (1) Yes (2) No If yes how many?<br />

11.Is the business service-oriented or goods- production enterprise? Service-<br />

oriented (1) goods-production enterprise (2)<br />

12. How many years has it been in existence? Less than 2 years (1) 2-5 years<br />

(2) 5-10 years ((3) Over 10 years (4)<br />

13. Are there many other such business enterprises in the neighbourhood? Yes (1) No (2)<br />

14. If yes how many? (1) 1-5 (2) 6-10 (3) Over 10.<br />

15. What facilities do they share if any?__________________________________<br />

Section C: Use of ICT (Information and Communication Technologies like the phone, computer<br />

and internet) for the business<br />

16. Does your business use ICTs? (1) Yes (2) No. If Yes, for what are the<br />

ICTs used for (1) To get information on raw materials/ordering goods (2) To get<br />

information on Customers, markets for goods/services (3) To contact friends and<br />

relatives (4). Any other need, specify _________________________ _<br />

17. How do you get information for your business? (1) Personal knowledge (2) Fellow<br />

Traders/producers (3) Mass media (radio, TV, newspapers) (4) internet (5) mobile<br />

phone contacts (4) Government officers (5) Acquaintances, friends and relatives,<br />

Other sources specify__________________________________________________<br />

18. How do you get information about customers for your work? Customers<br />

come on their own (1) Word of mouth (2) Advertising in the mass media (radio, TV,<br />

newspapers) (3) Through the internet and email contacts (4) Mobile phone contacts<br />

19. What else and by whom do you think needs to be done to make your access to<br />

information through ICTs better? (1) government (2) Phone companies (3) Other,<br />

specify ______________________________________________________________<br />

Section D :Acquisition and Use of (ICTs)<br />

518


Wakari Gikenye<br />

20. Does your business own a computer\mobile phone /or land line? Or other ICT<br />

resource, Specify_______________________________________________________<br />

21. How is/are the ICT resources used in the enterprise? (1) Typing<br />

(2) Communication (3) Internet use. (4) scanning (5) fax services (6) Sending money<br />

22. Do you think ICT resources are useful to the enterprise? Yes (1) No (2) Give<br />

Reasons _____________________________________________________________<br />

____________________________________________________________________<br />

23. Are there any people with computer skills in your organisation? (1) Yes (2) No<br />

24. If yes how many? (1) only one (2) all employees<br />

25. What kind of skill do they possess? (1) typing skills (2) skill for email and internet<br />

use (3) technical skills<br />

26. Are these skills used for the organisation? (1) Yes (2) No<br />

27. If yes what are the skills used for? (1) Typing (2) emails for keeping in touch with<br />

friends and relatives (3) business communication.<br />

28. What benefits does the use of ICTs bring to the business enterprise? ______________<br />

_____________________________________________________________________<br />

29. Do you know what was the approximate capital outlay at the beginning of this<br />

business? (1) Kshs 10,000-20, 000 (2) 21,000-40,000 (3) 41,000-60,000 (4) 61,000-<br />

100,000 (5) Over 100,000, Specify (6) Do not know.<br />

30. What is your approximate monthly turn-over (1) Kshs 10,000-20,000 (2) 21,000-<br />

40,000 (3) 41,000-60,000 (4) 61,000-100,000 (5) Other specify (6) Do not know.<br />

Section E: Access, Use and Impact of Mobile Phones<br />

31.How long have you had your mobile phone? (1) 1-2 years (2) 2-5 years (3) 5-10 years<br />

32. What do you use your mobile phone for? (1) For my work (2) For ordering goods (3)<br />

For contacting customers (4) For improving my business (5) For social<br />

communication (6) Other, specify ________________________________________<br />

33. How can you compare the time you did not have a mobile phone and now ?<br />

34. What other uses do you use your mobile phone for (1) For telling time (2) as a camera<br />

(3) for internet (4) For Mpesa (5) As a radio<br />

35. Has the mobile phone helped you to improve your business? (1) Yes (2) No<br />

36. What are the challenges\problems of using the mobile phone? (1) expensive to buy<br />

(2) air time (3) network congestion (4) easy target for thieves and pickpockets.<br />

37. Do you make/receive more business /more social calls on you mobile phone?<br />

(1) More Business-related calls (2) More social calls<br />

38. How many of the calls that you have made for the last one week are (1) work related<br />

(2) For social communication.<br />

39. Those which are work-related what were they used for? (1) To contact customers<br />

(2) To order goods\supplies (3) Other, Specify_____________________________<br />

40. Are you a registered Mpesa user? (1) Yes (2) No<br />

519


Wakari Gikenye<br />

41. How often do you use Mpesa (1) Everyday (2) Once a week (3) Other specify______<br />

42. What do you use Mpesa for? (1) T o send\receive money to\from relatives (2) For<br />

Business transactions.<br />

References<br />

Camner G. et al (2009). What makes a Successful Mobile Money Implementation? M-pesa in Kenya and Tanzania.<br />

Financial sector Deepening Kenya<br />

Communication Commision of Kenya (CCK) Quarterly sector statistics Report 2 nd quarter October 2009 –March2010<br />

http://wwwcck.go.ke./news/2010/news-30mar10.html<br />

Daily Nation, Wednesday September 29 2010 Mobile Banking boosts access to cash services. Nation Media group<br />

Limited.Nairobi.<br />

Geach N (2007) The Digital Divide, Financial exclusion and Mobile Phone Technology:Two Problems One Solution?<br />

British and Irish Low, Education and Technology Association 2007 Annual <strong>Conference</strong>. Hertfordshire16-<br />

17April. International Telecommunications Union (ITU) 2009. Measuring the Information society The ICT<br />

DevelopmentIndex.<br />

Kalba K. (2008) The adoption of mobile phones in emerging markets. Global diffusion and the rural challenge.<br />

International journal of Communication 2 (2008), P631-661.<br />

Khasawneh M. and Ibrahim H. (2008) Toward an information and communication technology development in<br />

developing countries. Communications of the International Business Information Management Association<br />

(IBIMA) Vol. 4 2008 p.135-140.<br />

Kimenyi M. and Ndungu N. (2009). Expanding the financial services frontier : Lessons from mobile phone banking in<br />

Kenya. Brookings Washington D.C.<br />

Mas I.and Morawczynski (2009) Designing mobile money services: Lessons from M-PESA. Innovations, a quarterly<br />

journal published by MIT Press vol.4, issue 2 2009 p. 77-91.<br />

Mas I and Radcliffe D. (2010). Mobile payments go viral : M-pesa in Kenya. Bill and Melinda Gates Foundation.<br />

Njenga A. n.d. Mobile phone banking: Usage experiences in Kenya. Catholic University of Eastern Africa.<br />

Okello-Obura, C. et al (2007). Assessment of business information access problems in Uganda. Partnership: The<br />

Canadian journal of library and information practice and research, vol.2 no.2 2007 pp.1-34.<br />

Omwansa T. (2009). M-pesa : Progress and prospects. Innovations/ Mobile World Congress 2009 p. 107-123<br />

Rogers E. (2003) Diffusion of innovations (5 th ed) New York Free Press.<br />

Standard Chartered, Asia, Africa and the Middle East : The guide to working capital Management 2009/2010.Banking<br />

the unbanked : Going mobile in Africa.<br />

M-pesa launched in South Africa (Vodacom) http:www.howwemadeitinafrica.com/M-pesa-launched-in-southafrica/3611/<br />

accessed on 22 nd September 2010.<br />

520


Querying DTI Analysis Results Within Deep Perisylvian Area<br />

Sarmad Istephan and Mohammad Siadat<br />

Oakland University, Rochester, USA<br />

sistephan@gmail.com<br />

siadat@oakland.edu<br />

Abstract: This paper is an advancement in the field of Data Driven Medicine. There has been a tremendous amount<br />

of work done in integrating technology into the medical field. Part of that work is in the storage and querying of<br />

medical images which is the focus of this paper. The high level motivation of this paper is to design a database<br />

driven system that empowers radiologists to validate their hypotheses against a set of medical images. More<br />

specifically, a system has been developed that answers queries pertaining to the average Fractional Anisotropy (FA)<br />

within a given brain structure. The proposed system enables querying of Diffusion Tensor Imaging (DTI) images to<br />

evaluate hypotheses that correlate changes in DTI analysis results (e.g., FA map) with brain epileptogenicity. The<br />

system makes the assumption that the following data are available: a) high-resolution and high-contrast image data<br />

such as T1-weighted MRI, b) models of brain structures of interest segmented in the high-resolution and highcontrast<br />

image space, c) DTI analysis maps such as FA map, d) registration Information that transfers the brain<br />

structures’ models from their native image space (T1) to target image space (FA map). A data model to store this<br />

information has been designed. The data model is implemented in a database that uses SQL Server 2008<br />

technology. This database stores patients’ data, various image modalities, segmented models, and registration<br />

information. Part of importing the images into the database includes reading the dicom image tags and feeding the<br />

available tags into the database. After all the required pieces of data are in the database, the calculation of the<br />

average FA within a given model is done as follows. First, using the Visualization Tool Kit (VTK) libraries and the<br />

registration information the segmented model is transferred from its native space to the target image space (e.g.,<br />

from T1-weighted MRI to FA map). Next, a bonding box is created around the model and each voxel within the<br />

bonding box is examined to see if it is inside the segmented model. At this point all the voxels of interest (inside the<br />

model) are identified and the average FA map would be calculated using the Image Tool Kit (ITK) libraries. The<br />

primary focus of this work was to calculate the average FA but the system can be extended to include many other<br />

features of interest such as fiber density, curvature measures, etc. In conclusion, this system creates a link between<br />

quantitative measurements of several modalities of data e.g. DTI and qualitative patients’ situations in clinical<br />

practices, which is an important element in the field of Data Driven Medicine.<br />

Keywords: medical image querying, medical image database, DTI/MRI querying, medical image data model,<br />

DICOM, medical imaging<br />

1. Introduction<br />

Data Driven Medicine is an initiative to improve health care in the United States by leveraging technology.<br />

The United States Department of Health and Human Services (HHS) has recently received a $3.6 billion<br />

budget to improve Information Technology (IT) within health care. A part of that budget goes to Health IT<br />

which is part of HHS. Health IT replaces traditional paper methods for storing patient information with<br />

Electronic Health Records (EHRs). The usage of EHRs enables the storage and secure dissemination of<br />

comprehensive patients’ health history. The availability of EHRs provides doctors with a mechanism to<br />

reach an accurate diagnosis sooner than traditional paper methods would (HHS Press Office 2010; The<br />

Office of the National Coordinator for Health Information Technology 2010).<br />

Patients’ medical image scans of various modalities are a subset of EHRs. These modalities include<br />

Magnetic Resonance Images (MRI), Diffusion Tensor Images (DTI), Computed Tomography (CT) and<br />

Ultra Sound (US). There are many medical image modalities created daily. In 2002, on a daily basis, at<br />

least 12,000 radiological images were created at the University Hospital of Geneva. In one year this<br />

amounts to 1 terabyte of image data (Muller et al 2004). Another study showed that tens of terabytes of<br />

image data is produced by an average size radiology department which means pedabytes of image data<br />

per country per year (Glatard et al Magnin 2004). This humongous amount of data can be used to<br />

complement traditional medicine in many ways. It is impossible to ask radiologists to analyze this data<br />

because their time is expensive and analyzing this amount of data manually would take a very long time.<br />

Therefore, a system must be designed that can automatically query the data.<br />

This paper focuses on developing a system with a database to store medical image modalities and<br />

develops a mechanism to query the image data automatically. Since this is an automatic system, then it<br />

addresses subjective analysis that may occur if a radiologist was performing the analysis. Also, the<br />

system will be able to analyze many more aspects of the images much faster than a radiologist can.<br />

521


Sarmad Istephan and Mohammad Siadat<br />

A data driven system has potential to be used in many areas. One usage of such a system is in medical<br />

school education where students can leverage the system to construct and test various hypotheses that<br />

they develop and verify them across many images. A second use of this system is to leverage the image<br />

data to determine an accurate quantitative outcome prediction. A quantitative outcome prediction would<br />

provide the patient with information using their image modalities in conjunction with those from previous<br />

patients with a similar disease to come up with conclusions. An example of a conclusion would be that<br />

90% of patients going through treatment plan X had side effect Y and it is up to the patient whether to<br />

undertake such a treatment that only 10% of patients did not have the side effect of. A third use of such a<br />

system is to come up with patient centric treatment plans. This means that treatment plan X can be<br />

tailored to a specific patient by leveraging information from the image data. A fourth use of such a system<br />

is as a decision support system. A decision support system will use the pre-operation and post-operation<br />

image data from previous patients during a new patient’s invasive surgery to decide on where to cut in<br />

order to minimize/eliminate tissue damage. Finally, the image data can be used as a mechanism to<br />

prevent the need for invasive techniques all together because it may unlock other treatment options that<br />

would not have been discovered by only using patient provided information (Muller et al 2004).<br />

The remainder of the paper has the following structure. In Section 2, a brief examination of previous work<br />

is provided. In Section 3, the method this paper uses is explained. In section 4, the technology used for is<br />

provided. In section 5, simulated data and results are provided. Finally, in section 6, the paper is<br />

concluded.<br />

2. Previous work<br />

There are three approaches to query medical data. One approach is to use the structured data.<br />

Structured data includes demographic data of a patient such as age, gender, height, weight, etc. Another<br />

approach is to use unstructured data. Unstructured data includes the image data of patients. The third<br />

approach is to use a hybrid of structured and unstructured data (Siadat et al 2009). This paper focuses<br />

on developing hypotheses and validating them using the unstructured image data.<br />

There are two methods used to query unstructured image data. One method is to create structured data<br />

(annotations) that represents the unstructured data. This approach is accomplished manually by a<br />

human. Sine transcribing unstructured data into structured data is manual then it is prone to error in the<br />

sense that the human can make mistakes while dictating the data. Also, it is irreproducible in that two<br />

humans may come up with different annotations or even the same human might come up with different<br />

annotations at different times for the same image data. In addition, this manual process is time<br />

consuming because only one volumetric image can be operated on at a given time. Finally, it is neither<br />

feasible nor possible to have a complete set of structured data to describe all the features of an<br />

unstructured image (Chbeir et al 2002) (Glatard et al 2004).<br />

The other method to query unstructured data is through the use of image segmentation, feature<br />

extraction and other imaging techniques that operate directly on the image itself. This approach<br />

addresses all the concerns of using structured data to represent unstructured data and hence is the<br />

preferable approach (Chbeir et al 2002). This paper queries the unstructured data directly to answer<br />

queries.<br />

The hybrid system approach to querying medical data is to use the unstructured image data as a<br />

complement to traditional textual queries. Textual queries are those pertaining to structured data. The<br />

querying of structured data will always be needed as a way to focus the study on certain demographics.<br />

However, leveraging unstructured data as a complement to the structured data provides the ability to<br />

answer many questions that are difficult to answer otherwise (Muller et al 2004). An example of such<br />

questions is, “Is there an intensity difference in the medial side of the perisylvian area of the left<br />

hemisphere of the brain compared to the medial side of the perisylvian area in the right hemisphere for<br />

patients within the ages of 20 and 25?” The age part of the question uses structured data and everything<br />

before that uses the unstructured data. (Chbeir 2002).<br />

There have been many good systems designed so far that use unstructured data. Most of these systems<br />

are designed using the Content Based Image Retrieval (CBIR) technique. In this technique, a system is<br />

given an image and asked to find all similar images in the database. The process of finding the similar<br />

images in the database is then refined by selecting an image from the result set and asking the system to<br />

again search. This iterative process is repeated over and over until accurate results are reached. The<br />

Spine and Pathology and Image Retrieval System (SPIRS) achieved a 20% increase after three iterations<br />

522


Sarmad Istephan and Mohammad Siadat<br />

of similarity based retrieval (Hsu et al 2008). Another system that uses CBIR is the Medical Image<br />

Management System (MIMS) which combines the use of structured and unstructured data to answer<br />

queries (Chbeir 2002). Another system that uses CBIR by exposing various methods as Web Services is<br />

MedImGrid (Jin 2007). Another system that performs Gabor Filter feature extraction and CBIR is done by<br />

Tristan Glatard (Glatard 2004). There are many other papers that also use CBIR as a mechanism to<br />

query unstructured data and are all reviewed in Henning Muller’s review of CBIR systems (Muller 2004).<br />

The theme amongst all these systems is the requirement of an initial image to initiate a query. This paper<br />

develops a system that does not require an initial image. Furthermore, these systems compare against<br />

one image modality. However, the system in this paper is able to do registration between different<br />

modalities of images thus allowing registering anatomical information to functional information.<br />

The research in medical data storage and retrieval has resulted in many systems that are widely used in<br />

hospitals. The overall system is the Hospital Information Systems (HIS). The HIS system is a system that<br />

consists of many subsystems used in the capturing, storage and retrieval of everything from patient<br />

appointments to image modalities acquisition to billing (Biohealthmatics 06). One subsystem of HIS is the<br />

Picture Archiving and communication Systems (PACS). PACS allows for the capturing, storing and<br />

retrieval of medical image modalities without the use of traditional image films (Becker and Arenson 94).<br />

A complement to PACS is the Radiological Information Systems (RIS). RIS enables the storage of patient<br />

data such as patients’ history from admission to discharge, patients’ appointments, patient centric reports<br />

and traditional film tracking (Biohealthmatics 06). The future of the system designed in this paper is a<br />

complement to all of these systems with the primary focus on the validation of hypotheses by querying<br />

the unstructured image data.<br />

3. Method<br />

In this paper, a new system was developed that enables the storage of unstructured medical images and<br />

hypotheses validation through querying the images directly. The goal of the system is to extend the<br />

power of Structured Query Language (SQL) through the use of the Common Library Runtime (CLR) to<br />

query the unstructured data. The general form of the query the system supports is,<br />

SELECT [Field Name]<br />

FROM [Database.Tables]<br />

WHERE [FeatureX] [Spatial Relation to Structure Name] [Structure Name] of [Image Map] [Comparison<br />

Operator] [Comparison Value]<br />

An example of this query is:<br />

SELECT [Field Name = Patient Name]<br />

FROM [Database.Tables]<br />

WHERE [FeatureX = AverageFA] [Spatial Relation to Structure Name = Adjacent to Medial Side] of<br />

[Structure Name = Perisylvian Area] of [Image Map = DTI] [Comparison Operator = >] [Comparison Value<br />

= 50]<br />

This general query has been encapsulated into a function called QueryImage. QueryImage function is a<br />

SQL CLR function that can be called directly from a query’s where clause. This function was created as<br />

an easy mechanism for the communication between the application and the database.<br />

QueryImage takes the following parameters,<br />

[FeatureX]: Type of module to be executed (e.g., Average FA).<br />

[Spatial Relation to Structure Name]: Geometric relationship to be used (e.g., within).<br />

[Structure Name]: The name of the structured to be queried (e.g., Perisylvian area).<br />

[Image Map]: The type of image modality to be used (e.g., MRI).<br />

[Comparison Operator]: The type of comparison to make (e.g., >).<br />

[Comparison Value]: The value to compare against (e.g. 50)<br />

523


Sarmad Istephan and Mohammad Siadat<br />

The power of this design is that new modules can be added to the system at any time without affecting<br />

the rest of the system. Using the AverageFA module, a radiologist can form the hypothesis, “the Average<br />

FA goes down next to an epileptogenic area.” This hypothesis can be answered using the query,<br />

SELECT PatientName<br />

FROM [Database.Tables]<br />

WHERE QueryImage(“AverageFA”, “ Adjacent to Medial Side”, “Perisylvian Area”, “DTI”, “50). The system then fetches the next patient and the cycle repeats itself.<br />

524


No<br />

No<br />

No<br />

Execute Query<br />

Get Next Patient<br />

Is there more<br />

patients<br />

Yes<br />

Is FeatureX<br />

Available?<br />

No<br />

Is Image Map<br />

available?<br />

Yes<br />

Is Model<br />

available?<br />

Yes<br />

Is Registration<br />

Information<br />

available?<br />

Sarmad Istephan and Mohammad Siadat<br />

Yes<br />

Yes<br />

Figure 2: System’s query execution architecture<br />

No<br />

Database<br />

Return Result Set<br />

Generate<br />

SpacialRelationToStructure<br />

Name<br />

Transfer<br />

SpatialRelationOfInterest<br />

model to ImageMap<br />

Calculate FeatureX<br />

Save feature value<br />

Is feature true for<br />

ComaprisonValue Using<br />

ComaprisonOperator?<br />

Yes<br />

Accumulate result<br />

into Result Set<br />

The steps to generate spatial relation to the structure name, the transfer of the relation and the<br />

calculation of feature are depicted in figure 3. This is done by the application first receiving the query,<br />

getting the image map, model and registration information from the database and calculating featurex<br />

using VTK libraries and Image Tool Kit (ITK) libraries. More specifically, VTK library is used for the<br />

following steps. First, load the model data from the database. Once the VTK polydata model is loaded<br />

then transfer the model from the model space to the image space using the registration information. The<br />

transfer of the model to the image space is illustrated in figure 4. Next, a bounding box is drawn around<br />

the model. Then an array containing a 1 for inside and 0 for outside for each voxel from the bounding box<br />

is generated. Figure 5 shows two voxels, once outside the model and one inside the model. Then, ITK<br />

libraries are used to load the image space volume. Finally, the sum of the intensity values for all inside<br />

voxels from the image space are all summed and divided by the total number of inside voxels. This sum<br />

represents the average FA.<br />

525<br />

No


Sarmad Istephan and Mohammad Siadat<br />

Figure 3: Details of the steps generation of the spatial relation to the structure name, the transfer of the<br />

relation and the calculation of feature from Figure 2<br />

Figure 4: The model space transformed to Figure 5: Model showing voxel outside and inside<br />

the image space<br />

4. Technologies used<br />

The data model is implemented in a database that uses SQL Server 2008 technology. This database<br />

stores patient data, various image modalities, registration information and segmented model information.<br />

An ASP.NET web application has been built on top of this database leveraging the .NET Framework 4.0<br />

technology. This web application contains forms in which the available images are imported into the<br />

database. The backend for the forms is written using the C# programming language. Furthermore, the<br />

backend code uses ClearCanvas to read the dicom image tags. In addition, C# VTK library DLLs and C#<br />

ITK library DLLs are used form image and model manipulation. It should be noted that both VTK library<br />

and the ITK library are written in C++ and therefore, CMAKE was used to create the C# DLLs from the<br />

native C++ code. Finally, Matlab was used to generate simulated image volumes and 3D Slicer was used<br />

to generate VTK models for the image data.<br />

5. Simulated data and result<br />

In this system, simulated data was generated and used to test the system because the outcome is<br />

known. It would be difficult to use real MRI/DTI data to verify the system because the outcome is not<br />

known. The simulated data includes sets consisting of a DICOM volume (see figure 8 for an example), a<br />

surface model (see figure 9 for an example) and registration information. The system was verified that it<br />

is able to query the simulated images using the QueryImage function.<br />

The process of generating the simulated data is depicted in figure 6. First, Matlab was used to randomly<br />

generate the model’s parameters. R1, R2, Zmin, Zmax, Xmin, Xmax, Ymin and Ymax. The R1 represents<br />

the radius from the center of the image to the outer edge of the structure. R2 represents the radius from<br />

the center of the image to the inner edge of the structure. In other words, R1 and R2 are used to control<br />

the width of the generated structure. Zmin and Zmax is the minimum and maximum number of slices the<br />

system is to generate (see figure 7). Xmin, Xmax and Ymin, Ymax are the dimensions of the slices. Using<br />

526


Sarmad Istephan and Mohammad Siadat<br />

the values of R1, R2, Zmin, Zmax, Xmin, Xmax, Ymin and Ymax, a model is generated and then saved<br />

as a DICOM image (see figure 8). If the system is to generate more images then another set of<br />

parameters are evaluated creating another DICOM image.<br />

The models are the imported into 3D slicer to generate a geometrical surface model which is then saved<br />

in VTK format. Then, the model is loaded using VTK libraries and a transformation (translation and<br />

rotation) is applied to it. Finally, all this information is saved into the database using the web forms. At this<br />

point the system is ready to be queried.<br />

Figure 6: Simulated data generation<br />

Figure 7: Axial, Coronal and Sagittal views of a slice within the 3 volume<br />

Figure 8: Model created out of the 3D volume<br />

6. Conclusion<br />

This paper created a system that is used to store and query medical images which serves as an<br />

extension to the field of Data Driven Medicine. This system can be used in academia, hospital research<br />

and hospital clinical diagnosis and prognosis. This paper shows the feasibility of such a system and will<br />

be extended into a complete system in future work by adding more module types such as fiber density<br />

measures, curvature measures and standard deviation of the intensity modules.<br />

References<br />

Becker, S. and Arenson Ronald (1994) “Costs and Benefits of Picture Archiving and Communication Systems”,<br />

Journal of the American Medical Informatics Association, vol. 1:4, November, pp.361 – 371.<br />

527


Sarmad Istephan and Mohammad Siadat<br />

Biohealthmatics (2006), Healthcare Technologies [Online], Available:<br />

http://www.biohealthmatics.com/technologies/healthtech.aspx [3 Dec 2010].<br />

Chbeir, R., Atnafu, S. and Brunie, L. (2002) “Image Data Model for an Efficient Multi-Criteria Query: A Case in<br />

Medical Databases”, Proceedings of the 14th International <strong>Conference</strong> on Scientific and Statistical Database<br />

Management IEEE. pp. 165-174.<br />

Glatard, T., Montagnat, J. and Magnin, I. (2004) “Texture Based Medical Image Indexing and Retrieval: Application to<br />

Cardiac Imaging”, Proceedings of the <strong>6th</strong> ACM SIGMM international workshop on Multimedia information<br />

retrieval.<br />

HHS Press Office. (2010) HHS Budget Makes Smart Investments, Protects the Health and Safety of America’s<br />

Families, [Online], Available: http://www.hhs.gov/news/press/2010pres/02/20100201a.html [29 Nov 2010].<br />

Hsu, W., Antani, S., Long, R. Neve, L. and Thoma, G. (2009) “SPIRS: A Web-based image retrieval system for large<br />

biomedical databases”, International journal of medical informatics, vol 78, pp. 13-24.<br />

Jin, H., Sun, A., Zheng, R. He, R. Zhang, Q., Shi, Y., Yang, W. (2007) “Content and Semantic Context Based Image<br />

Retrieval for Medical Image Grid”, GRID '07 Proceedings of the 8th IEEE/ACM International <strong>Conference</strong> on Grid<br />

Computing.<br />

Muller, H., Michoux, N., Bandon D. and Geissbuhler, A. (2004) “A review of content-based image retrieval systems in<br />

medical applications - clinical benefits and future directions”, International Journal of Medical Informatics, vol.<br />

73, pp 1-23.<br />

Siadat, M., Hammad, R., Shetty, A., Soltanian-Zadeh, H., Sethi, I., Eetemadi, A., Elisvich, K. (2009) “DTI data<br />

modeling for unlimited query support”, SPIE, Vol. 7264.<br />

The Office of the National Coordinator for Health Information Technology (2010), Why Health IT? [Online], Available:<br />

http://healthit.hhs.gov/portal/server.pt/community/healthit_hhs_gov__home/1204 [29 Nov 2010].<br />

528


Towards an Evaluation Method for Information Quality<br />

Management of Health Information Systems<br />

Siti Asma Mohammed and Maryati Mohd Yusof<br />

National University of Malaysia, Bangi, Selangor<br />

asmadixie@gmail.com<br />

mmy@ftsm.ukm.my<br />

Abstract: Information quality has been regarded by most researchers as one of the critical success factors of Health<br />

Information Systems (HIS) adoption. However, there is increasing evidence of poor quality information found in HIS.<br />

Poor information quality increases medical errors and adverse events. The occurrence of poor information quality<br />

reveals the need to evaluate information quality management (IQM) practices in healthcare organizations. It is crucial<br />

to assess the effectiveness of IQM to ensure the quality of information produced by HIS. A review of current<br />

evaluation methods for IQM in healthcare has been carried out to identify any existing gaps for further improvement.<br />

The literature review was conducted from the health informatics and information systems (IS) disciplines. It was<br />

found that the lack of information quality was due to a combination of human, organizational and technological<br />

factors. In the literature, guidelines and best practices are suggested in order to manage information quality in<br />

healthcare. However, there is inadequate literature that offers specific approaches and criteria for assessing quality<br />

information from the perspective of humans, organizations and technology. Therefore, this paper proposes a<br />

comprehensive and systematic evaluation method for HIS information quality management in improving the quality of<br />

information in HIS. The proposed method is developed based on PMBOK (Project Management Body of Knowledge)<br />

in order to provide a systematic evaluation process. With regards to evaluation criteria, this research focuses<br />

primarily on the criteria identified from a combination of quality management, IQM and IS/HIS evaluation findings<br />

from the literature. This paper contributes to the information quality body of knowledge within the information systems<br />

area. This evaluation method can be applied as a tool to evaluate the IQM practice in healthcare organizations and<br />

assist HIS users in improving the quality of information.<br />

Keywords: health information systems, information quality management, evaluation, project management, human<br />

factors, organizational factors<br />

1. Introduction<br />

Most researchers consider information quality (IQ) as one of the critical factors for the success of HIS<br />

(DeLone and McLean, 1992; Heeks, 2006; Yusof et al, 2008). Unfortunately, increasing evidence shows<br />

that HIS has not been producing the intended quality of information (Gaikwad et al, 2007; Miettinen and<br />

Korhonen, 2008; Agnew-Blais et al, 2009). Poor IQ increases medical errors and affects the quality of<br />

patient care (Institute of Medicine, 2000). Human, organizational and technological factors are among the<br />

elements that influence poor IQ in HIS (Caccia-Bava, Guimaraes and Harrington, 2006; Kaplan and<br />

Harris-Salamone, 2009). For example, organizational studies show that users are unable to align<br />

information usage with management and clinical performance (Garrib et al, 2008). Further, poor IQ<br />

occurs because of unclear responsibility for IQ and lack of clarity in IQ expectations among users (Kerr,<br />

2006). Lack of quality requirement specification is the reason why users have little understanding of the<br />

purpose of collecting and utilizing information to support decision making (Sutherland and Steinum,<br />

2009). Consequently, users consider less to produce quality information, which later becomes the main<br />

hindrance to adopting HIS (Orfanidis et al, 2004).<br />

2. Theoretical background<br />

Having Information quality management (IQM) practices is necessary to achieve optimal quality of<br />

information. IQM is a set of activities where an organization establishes, controls, conducts quality<br />

strategies so that the process of producing information satisfies user requirements (Caballero et al,<br />

2008). These activities include establishment of roles, responsibilities, IQ requirements and procedures.<br />

IQM provides guidance on the meaning of IQ to the organization (Price and Shanks, 2005). Information is<br />

of high quality if it is fit for use where the information is able to satisfy the actual use of the information<br />

(Tayi and Ballou, 1998). IQ and data quality (DQ) are always used interchangeably in the literature by<br />

most researchers. Therefore, this paper will also use the term IQ interchangeably with DQ unless<br />

specified. There are guidelines and best practices being developed for having strategic IQM in healthcare<br />

organizations (AHIMA, 1998; Arts et al, 2002; Kerr, 2006; Aqil, Lippeveld and Hozumi, 2009). However,<br />

the most critical factors for IQM and the criteria are still unclear to benchmark the quality of HIS<br />

information (Ammenwerth, Breu and Paech, 2010). In addition, more research is needed to study the<br />

practicalities of information use such as the relationships between users and the HIS in disseminating<br />

information, the dynamic changes of clinical relationship in addressing information discrepancy and other<br />

529


Siti Asma Mohammed and Maryati Mohd Yusof<br />

stakeholders’ influences (Greenhalgh et al, 2009). Evaluation can be undertaken to investigate the IQM<br />

practice in healthcare organizations for effective HIS. A more holistic evaluation approach is needed to<br />

understand the production of IQ by HIS given human and organizational interventions. To date, there is<br />

inadequate literature that offers a comprehensive and systematic evaluation approach together with<br />

specific criteria for having quality information. Hence, more work is required to investigate this issue<br />

further.<br />

2.1 Previous works on HIS information quality management<br />

Table 1 below shows the review of best practices that have been developed to strategically manage<br />

information production in HIS. This review attempts to identify the gap in HIS IQM.<br />

Table 1: Selected Works on HIS IQM<br />

Author Objective Information<br />

Life Cycle<br />

Factors and Criteria<br />

Method<br />

P A M A H O T<br />

Data Quality<br />

Assurance in<br />

Medical<br />

Registries<br />

(Arts et al,<br />

2002 )<br />

AHIMA Data<br />

Quality<br />

Management<br />

Model<br />

(AHIMA,<br />

1998)<br />

Data Quality<br />

Improvement<br />

Strategy<br />

(Kerr, 2006)<br />

PRISM<br />

(Aqil,<br />

Lippeveld<br />

and Hozumi,<br />

2009)<br />

1) To identify<br />

the type<br />

and causes of<br />

poor data<br />

quality<br />

2) To develop<br />

data quality<br />

assurance<br />

procedures for<br />

medical<br />

registries<br />

To ensure data<br />

quality<br />

and continuous<br />

quality<br />

improvement<br />

To<br />

institutionalize<br />

and establish<br />

strategic<br />

management of<br />

data quality in<br />

the New<br />

Zealand health<br />

sector<br />

To evaluate the<br />

performance of<br />

HIS<br />

from the<br />

perspective of<br />

information<br />

quality handling<br />

X X X Training<br />

Motivation<br />

Quality follow<br />

up<br />

X X X Data collection<br />

purpose<br />

Education<br />

Training<br />

Communicate<br />

changes and<br />

updates<br />

Accountability<br />

Quality data<br />

sources<br />

X X X Quality data<br />

source<br />

Training<br />

Quality<br />

awareness<br />

Peer review<br />

Frequent<br />

checks,<br />

analysis and<br />

profiling of<br />

data<br />

Data quality<br />

mapping<br />

(business<br />

rules and data<br />

quality<br />

metrics)<br />

X X X Checking skill<br />

HIS tasks<br />

competency<br />

Motivation<br />

Confidence for<br />

HIS tasks<br />

530<br />

Data<br />

requirements<br />

Quality<br />

assurance plan<br />

Ethics<br />

Scope<br />

Information<br />

requirements<br />

Processes<br />

Documentation<br />

Monitoring<br />

Regulation<br />

Data<br />

stewardship<br />

Governance<br />

Planning<br />

Resources<br />

Training<br />

Supervision<br />

Information<br />

quality<br />

culture<br />

promotion<br />

Database<br />

design and<br />

function<br />

Technical<br />

support<br />

Interface<br />

usability<br />

Data<br />

ownership<br />

Accessibility<br />

Correct<br />

algorithms<br />

and formulas<br />

Data<br />

currency<br />

Database<br />

normalization<br />

Database<br />

maintenance<br />

and support<br />

Database<br />

design and<br />

normalization<br />

Interface<br />

System<br />

maintenance<br />

Interface<br />

complexity<br />

HIS design<br />

IT complexity<br />

Formative<br />

Formative<br />

Formative<br />

Formative


Siti Asma Mohammed and Maryati Mohd Yusof<br />

Note: Information Life Cycle: P – Plan, A – Acquisition, M – Maintenance, A - Apply<br />

Factors: H – Human, O – Organization, T – Technology<br />

Generally, all the four methods mentioned in Table 1 provide clear guidance for having quality information<br />

in HIS. For example, Arts, Keizer and Scheffer developed a generic framework that outlines the<br />

procedures for data quality assurance for medical registry system. The procedures are divided into three<br />

categories; system registry setup and organisation (plan), detection during data collection (acquisition)<br />

and actions to be taken to improve data quality (maintain).<br />

As for American Health Information Management Association (AHIMA), an IQM model is developed to<br />

continuously improve IQ. The model includes data application (purpose for which data are collected),<br />

data collection (process by which data elements are accumulated), warehousing (processes and systems<br />

used to archive data) and analysis (process of translating data).<br />

Another IQM initiative, which is a DQ improvement strategy, has been developed by Kerr to promote DQ<br />

in New Zealand. This strategy was designed at the Ministry of Health New Zealand with involvement from<br />

healthcare stakeholders in New Zealand. This program is meant to provide DQ standard operating<br />

procedures for healthcare organizations in New Zealand.<br />

Aqil, Lippeveld and Hozumi developed a PRISM model that details the determinants influencing HIS<br />

performance from the perspective of IQ handling. The behavioral factors, which explain the human<br />

behavior in handling data, discussed briefly on the use of information technology. For example, HIS tasks<br />

competency and the confidence level of using HIS.<br />

The similarity between these studies is the emphasis on IQ behavior or IQ handling, which comes under<br />

the human factor. Human factors such as training, motivation, education and quality awareness are<br />

among the important determinants that will influence IQ. DQ requirements, clinical process, management<br />

supervision, IQ culture and IQ governance are the examples for organization factors. For the technical<br />

aspects, while AHIMA focusing on the stability of database design to achieve IQ, the other three methods<br />

include the importance of HIS design such as interface usability. On the other hand, Aqil, Lippeveld and<br />

Hozumi depicts that information technology usage will also affect IQ.<br />

It can be seen that all methods share similar opinion on the importance of quality awareness, DQ<br />

requirements and IT technical support in ensuring IQ. Further, the methods are applied across the<br />

information life cycle (ILC) through a formative evaluation. This research believes more can be added to<br />

the evaluation criteria proposed within the human, organizational and technological factors. There should<br />

be more explicit evaluation categories and evaluation criteria that can be assigned to each ILC. Thus, a<br />

holistic evaluation method that includes criteria from human, organization and technology factors<br />

throughout the ILC would offer a better solution.<br />

2.2 Project Management Body of Knowledge (PMBOK)<br />

In designing a systematic evaluation approach, this research recognizes the importance of project<br />

management discipline. We believe project management is important when conducting an evaluation.<br />

This is because project management is based on planning, organizing and controlling each activity so<br />

that evaluation objective is achieved (Irani, 2010). As such, a project management discipline, Project<br />

Management Body of Knowledge (PMBOK), becomes the basis for developing the evaluation method<br />

(PMI, 2008). PMBOK provides common operational steps with clear inputs and outputs. PMBOK is<br />

recognized as a good practice and contributes to project success and organization value (Zwikael, 2009).<br />

PMBOK defines project management as ‘the application of skills, tools and techniques to project activities<br />

to meet project requirements’ (PMI, 2008:6). PMBOK includes a collection of processes to manage all<br />

types of projects. These process groups are performed sequentially and iteratively prior to completing the<br />

project. Each process group has its individual processes that are linked by their inputs and outputs. The<br />

output of one process becomes the input of another. Tools and techniques are the mechanisms applied<br />

to create the output.<br />

In addition, most studies in IS evaluation have failed to focus on understanding the ‘how’ aspect or the<br />

evaluation process itself (Tuten, 2009; Irani, 2010). This may explain why IS evaluation in healthcare<br />

531


Siti Asma Mohammed and Maryati Mohd Yusof<br />

organization is unclear and does not provide a strong basis for a good evaluation theory and an effective<br />

evaluation method (Ammenwerth et al, 2004).<br />

Thus, this research believes that PMBOK complements IS evaluation in the sense that PMBOK provides<br />

more guidance and details on how to plan, execute, manage and control the evaluation method. This<br />

research uses PMBOK to develop the evaluation method to support users in improving the quality of<br />

information in HIS continuously.<br />

3. Proposed evaluation method for HIS information quality management<br />

The proposed evaluation model (Figure 1) is developed based on PMBOK (PMI, 2008:39). It describes<br />

the process for conducting the evaluation. This is done by phases throughout the ILC. In order to find the<br />

specific criteria that determine the production of IQ in HIS, this research makes use most of the criteria<br />

identified from a combination of quality management, IQ and IS/HIS evaluation findings. Three major<br />

factors that influence the production of quality information are identified; human, organization and<br />

technology. Even though those criteria do not explicitly focus on IQ, they are all relevant and applicable in<br />

evaluating HIS IQM. Some criteria used are also based on IQM best practices given in the literature such<br />

as the information behavior criteria. Important criteria pertaining to the use of HIS and user satisfaction<br />

towards the system in use are also included after reviewing the literature from health informatics. The<br />

way a user behaves towards the information he possesses, the way he uses the system and his<br />

satisfaction towards the system used will determine the quality of information produced. This research<br />

makes use of the HOT-fit (Human, Organization and Technology-fit) framework in categorizing the factors<br />

and dimensions (Yusof et al, 2008). This is because HOT-fit gives explicit evaluation categories, and<br />

indicates clear relationships and definitions between the evaluation factors and dimensions. However, the<br />

detail criteria are the findings from literature reviews pertaining to effective IQM. Therefore, criteria used<br />

for this study are categorized under different suitable factors and dimensions.<br />

Figure 1: Evaluation method for HIS information quality management<br />

The model above describes the evaluation phases, key activities, ILC involved under each evaluation<br />

phase, input (evaluation resource), evaluation output (deliverables), criteria for each ILC involved and the<br />

methods used to conduct the evaluation. Each evaluation phase has its own activities, inputs/outputs,<br />

criteria and methods. The evaluation criteria are distributed according to different ILC being evaluated.<br />

Some criteria are repeatedly included in different ILCs depending on its relevance. Below are the<br />

definitions for ILC stages for this research.<br />

Table 3: ILC stages definitions (adapted from English, 1999)<br />

ILC Stage Definition<br />

Plan Plan for resources including information requirements, IQ definitions and metrics, process design,<br />

information modeling, database architecture, information resources, user roles, system usage<br />

preparation and method for collecting the information.<br />

Acquisition Process of collecting, verifying and storing the data/information into HIS.<br />

Maintenance Ensure that all resources continue to operate properly to make the information assessable and<br />

usable. These include updating, changing, cleansing and validating data and system maintenance.<br />

Apply Information is used to accomplish objectives.<br />

532


3.1 Evaluation criteria<br />

3.1.1 Human<br />

Siti Asma Mohammed and Maryati Mohd Yusof<br />

System Use refers to the frequency and breadth of HIS use (Yusof et al, 2008). System Use is<br />

associated with higher performance of quality of care (Poon et al, 2010). System Use includes the<br />

training provided for the HIS users in using the system, the level of system acceptance, the level of<br />

motivation to use the system and their knowledge of the system that is beyond their current job<br />

assignments (Yusof et al, 2008). User role is another criterion for System Use. Clear user roles when<br />

using the system determines the quality of information in HIS. User roles refer to data collectors, data<br />

custodians and data consumers (Kerr, 2006; Wang et al, 1998). Information Behavior is the HIS user’s<br />

effort in ensuring quality in collecting, creating and using the information. Information Behavior criteria<br />

include education and training on medical information, and quality awareness on the information<br />

possessed. The information must be patient-centred and should satisfy the actual use of data consumers<br />

(Macinati, 2008; Ahire, 1996).<br />

3.1.2 Organization<br />

The organization Structure of healthcare organization has a major influence on the production of HIS IQ.<br />

Structure includes leadership, physician involvement, IQ governance, IQ culture, IQ team and IQ<br />

requirements. Leadership indicates the role and commitment of the top management in motivating quality<br />

culture among the HIS users (Wang and Strong, 1996; English, 1999). IQ culture is a very important<br />

criterion that is emphasized by English (English, 1999). It refers to organization culture transformation<br />

that regards IQM activities as a habit and mind-set by applying continuous quality improvement<br />

principles. IQ team is the main entity that initiates and encourages IQM activities. IQ requirements are<br />

part of the IQM activities where IQ standards and definitions are developed during the ILC planning<br />

stage.<br />

3.1.3 Technology<br />

System Quality affects the production of IQ. Database design, ease of use, availability, usefulness and<br />

data security and are among the important criteria for System Quality (Ammenwerth, Breu and Paech,<br />

2010). Database design refers to the data inside the database that is efficiently organized for easy<br />

retrieval and reduces anomalies. Ease of use includes the convenient of using HIS such as the interface<br />

usability. Availability refers to the run time of the HIS for a desirable or expected length of time. Service<br />

Quality is the overall support received from the IT team within the healthcare organization and from the<br />

HIS vendor. Criteria for Service Quality are responsiveness, follow-up, assurance and empathy (Yusof et<br />

al, 2008)<br />

3.2 Evaluation phase<br />

3.2.1 Initiation<br />

Evaluation initiation is the process of initiating a new evaluation project (PMI, 2008:44). During the<br />

initiation phase, the stakeholders with the most knowledge of the clinical process, medical information<br />

being collected, the HIS in use and those responsible for maintaining the quality of information, are<br />

identified. The stakeholders include the quality team, medical officers, physicians, nurses and HIS users.<br />

The stakeholders must be those who interact and influence the overall outcome of the evaluation project.<br />

These people will become the input for this evaluation phase and throughout the evaluation project. An<br />

analysis of the stakeholders, to understand the degree of influential and impact to the evaluation, will be<br />

done using Social Network Analysis (SNA). During this phase, the evaluation goal and the project scope<br />

are explained. The evaluator informs the stakeholders of the goal, the reasons for conducting the<br />

evaluation and the significance of the evaluation that triggered the evaluation. All parties are expected to<br />

agree to the evaluation before the scope is further defined. A decision is made; whether to continue,<br />

delay or discontinue the evaluation. The evaluation scope defines the evaluation boundaries, which<br />

include the parts of the process and system that are to be studied. The time and resources needed to<br />

conduct the evaluation are also negotiated during this initiation phase. The deliverables for this phase are<br />

the Evaluation Charter and Stakeholder Analysis. The Evaluation Charter formally authorizes the<br />

evaluation project and documents initial requirements that satisfy the stakeholder’s needs and<br />

expectations. Table 4 explains the parameters for evaluation initiation phase.<br />

533


Table 4: Parameters for evaluation initiation phase<br />

3.2.2 Planning<br />

Siti Asma Mohammed and Maryati Mohd Yusof<br />

Key Activity Identify stakeholders<br />

Sub-activities 1. Identify stakeholders<br />

2. Brief evaluation goal<br />

3. Discuss and agree on evaluation scope<br />

ILC -<br />

Input Stakeholders<br />

Evaluation Criteria -<br />

Output Evaluation charter, Stakeholder analysis<br />

Method Interviews, SNA<br />

Evaluation planning is the process of designing the evaluation (PMI, 2008:46). During this phase, the<br />

evaluator attempts to learn about the current system, the clinical process, the environment supporting the<br />

clinical process and the IQ requirements. The evaluator will try to understand the problem domain in<br />

which the IQ problems exist and assesses the status of quality management effort concerning the<br />

planning and strategy for having quality information in HIS. The quality management team is expected to<br />

describe and guide the evaluator on the initiatives about IQ. The aim of the appraisal is to understand the<br />

IQM effort at the planning stage of the ILC. Further, the evaluator will identify HIS user roles involved in<br />

the different stages of ILC and their degree of dependencies using SNA. The key deliverables for this<br />

phase will be Evaluation Plan, Workflow Analysis and SNA. The Evaluation Plan defines how to execute<br />

the evaluation. Table 5 describes the parameters for evaluation planning phase.<br />

Table 5: Parameters for evaluation planning phase<br />

Key Activity Quality management environment analysis<br />

Sub-activities 1. Review of strategic and operative environment<br />

2. Understand and assess the status of quality management effort<br />

ILC Plan<br />

Input Evaluation Charter, stakeholders<br />

Evaluation Criteria<br />

Human Organization Technology<br />

System Use<br />

Training<br />

User role<br />

Information Behavior<br />

Education and training<br />

Structure<br />

Management leadership<br />

Physician involvement<br />

IQ governance<br />

IQ culture<br />

IQ team<br />

IQ requirements<br />

Output Evaluation plan, Workflow Analysis, SNA, IQ standards<br />

Method Interviews , Document Analysis, Observations<br />

System Quality<br />

Database design<br />

Ease of use<br />

Availability<br />

Usefulness<br />

Data security<br />

3.2.3 Execution<br />

Evaluation execution is the process of performing the evaluation at the operational level where medical<br />

information is collected, stored, maintained and applied (PMI, 2008:55). The aim of this assessment is to<br />

get an overview of information production during the delivery of care. The criteria for this phase will be<br />

used to guide the evaluation. During the assessment, any possible issues are monitored and captured.<br />

The problems underlying these issues are analyzed. Table 6, Table 7 and Table 8 describe the<br />

parameters for execution phase according to three different stages of ILC, which are acquisition,<br />

maintenance and apply.<br />

534


Siti Asma Mohammed and Maryati Mohd Yusof<br />

Table 6: Parameters for evaluation execution phase (ILC - acquisition)<br />

Key Activity Information production process and business impact assessment<br />

Sub-activities 1. Records quality activities<br />

2. Monitor and capture various possible issues<br />

3. Analyze problems<br />

4. Identify, update and refine variances from the evaluation project<br />

ILC Acquisition<br />

Input Evaluation plan, Workflow Analysis, SNA, Quality team, HIS users<br />

Evaluation Criteria<br />

Human Organization Technology<br />

System Use<br />

Acceptance<br />

System motivation<br />

Knowledge/skills<br />

Information Behavior<br />

Quality awareness<br />

Patient centered<br />

Structure<br />

Physician involvement<br />

IQ culture<br />

Output Evaluation results, cause-and-effect analysis<br />

System Quality<br />

Database design<br />

Ease of use<br />

Availability<br />

Usefulness<br />

Data security<br />

Service Quality<br />

Responsiveness<br />

Follow up<br />

Assurance<br />

Empathy<br />

Method Interviews, Observations, Trend analysis, Expert and user judgments<br />

Table 7: Parameters for evaluation execution phase (ILC - maintenance)<br />

Key Activity System maintenance assessment<br />

Sub-activities 1. Investigate maintenance efforts<br />

2. Monitor and capture various possible issues<br />

3. Analyze problems<br />

4. Identify, update and refine variances from the evaluation project<br />

ILC Maintenance<br />

Input Evaluation plan, Workflow Analysis, Quality team, HIS users<br />

Evaluation Criteria<br />

Human Technology<br />

Information Behavior<br />

Quality awareness<br />

Motivation<br />

Patient centered<br />

Output Evaluation results, cause-and-effect analysis<br />

System Quality<br />

Database design<br />

Ease of use<br />

Availability<br />

Usefulness<br />

Data security<br />

Service Quality<br />

Responsiveness<br />

Follow-up<br />

Assurance<br />

Empathy team<br />

Vendor<br />

Method Interviews, Observations, Trend analysis, Expert and user judgments<br />

535


Siti Asma Mohammed and Maryati Mohd Yusof<br />

Table 8: Parameters for evaluation execution phase (ILC - apply)<br />

Key Activity IQ and business impact assessment<br />

Sub-activities 1. Assess and map the IQ usefulness based on users’ predefined IQ standards at planning<br />

stage.<br />

2. Assess the information satisfaction<br />

3. Identify, update and refine variances from the evaluation project<br />

ILC Apply<br />

Input Evaluation project plan, Workflow Analysis, Quality team, HIS users<br />

Evaluation<br />

Human<br />

Criteria Information Behavior<br />

Information satisfaction<br />

Output Evaluation results, cause-and-effect analysis<br />

Method Interviews, Observations, Trend analysis, Expert and user judgments<br />

3.2.4 Monitor and control<br />

Monitoring and controlling the evaluation occurs at every evaluation phase especially during the biggest<br />

part of evaluation which is the execution phase (PMI, 2008:59). Monitoring is to track the evaluation<br />

progress and to identify any variances or issues during evaluation. If necessary, the evaluator may<br />

control the issues by adjusting the scope, plan or resources. A progress report is necessary to gain an<br />

insight into the evaluation progress and to identify areas that need attention.<br />

3.2.5 Closure and actions<br />

The closure phase consists of activities to finalize and complete the evaluation project (PMI, 2008:64).<br />

During this phase, the evaluator communicates the evaluation findings and recommendations to the<br />

stakeholders. Experiences and lessons learned during evaluation are shared with the stakeholders.<br />

Based on the findings and lessons learned, this phase should be a continuous process of improvement.<br />

Future actions must be considered and implemented. An evaluation plan must be included to assess<br />

these future actions in order to continuously improve IQM practices in healthcare organization.<br />

4. Conclusions<br />

This paper has highlighted the issues of poor IQ in HIS, reviewed the existing works on IQM in healthcare<br />

organizations and proposed an evaluation method for IQM evaluation. From the literature, we found that<br />

poor IQ in HIS is influenced by three factors; human, organizational and technological factors. Review of<br />

current IQM practices in healthcare organization suggests that a comprehensive criteria for having quality<br />

information in HIS is needed. In addition, there is a need for a systematic and effective evaluation<br />

method. Thus, an evaluation method for IQM of HIS that is based on project management discipline is<br />

proposed. The evaluation method is designed based on PMBOK, to complements the method for<br />

conducting IS evaluation. In this paper, we have justified the need for project management knowledge in<br />

IS evaluation. The evaluation method also includes the evaluation criteria based on human,<br />

organizational and technological factors, which are assigned to each ILC according to its relevance. The<br />

evaluation method is meant to evaluate the IQM practices for HIS at the operational level.<br />

We believe this research contributes to IS discipline especially related to IQ and HIS evaluation. First, this<br />

research extends the knowledge that IQ is also influenced by the user’s behavior towards the system in<br />

use. Second, this research takes a holistic approach to examine the production of IQ in healthcare<br />

organization, which goes beyond usual studies that evaluate IQ to validate the HIS efficiency in producing<br />

quality information without taking into consideration the factors that affect quality. This research<br />

addressed the need for a better understanding of the HIS evaluation process by adapting project<br />

management discipline. Lastly, this research proposes a systematic evaluation method for HIS<br />

information quality management to support the users in improving continuously the quality of information<br />

in the HIS. For future work, opinions from experts are required to verify the criteria for having quality<br />

information and the evaluation method. Afterward, the evaluation method will be validated in healthcare<br />

organization for further refinement. This research will use qualitative study to gain an in-depth knowledge<br />

about HIS IQM in healthcare organization, with an intention to develop directions or predictions about the<br />

practice of HIS users in producing quality information.<br />

536


References<br />

Siti Asma Mohammed and Maryati Mohd Yusof<br />

Agnew-Blais, J. C., Coblyn, J. S., Katz, J. N., Anderson, R. J., Mehta, J. & Solomon, D. H. (2009) "Measuring Quality<br />

of care for rheumatic diseases using an electronic medical record". Annals Rheumatic Diseases, vol. 68, no. 5,<br />

May, pp. 680-684.<br />

AHIMA (1998) "Data Quality Management Model". Journal of AHIMA, vol 69, no. 6, June, pp 2-7.<br />

Ahire, S. L., Golhar, D. Y. & Waller, M. A. (1996) "Development and validation of TQM implementation constructs".<br />

Decision Sciences, vol. 27, no. 1, March, pp. 23-56.<br />

Ammenwerth, E., Brender, J., Nykanen, P., Prokosch, H.-U., Rigby, M. & Talmon, J. (2004) "Visions and strategies to<br />

improve evaluation of health information systems Reflections and lessons based on the HIS-EVAL workshop in<br />

Innsbruck". International Journal of Medical Informatics, vol. 73, no. 6, June, pp. 479-491.<br />

Ammenwerth, E., Breu, R. & Paech, B. (2010) User-Oriented Quality Assessment of IT-Supported Healthcare<br />

Processes - A Position Paper. In Rinderle-Ma, S., Sadiq, S. & Leymann, F. (Eds.). Business Process<br />

Management Workshop 2009. Springer Berlin Heidelberg.<br />

Ammenwerth, E., Ehlers, F., Hirsch, B. & Gratl, G. (2007) "HIS-Monitor: An approach to assess the quality of<br />

information processing in hospitals". International Journal of Medical Informatics, vol. 76, no. 2-3, February, pp<br />

216-225.<br />

Aqil, A., Lippeveld, T. & Hozumi, D. (2009) "PRISM framework: a paradigm shift for designing, strengthening and<br />

evaluating routine health information systems ". Health Policy and Planning, vol. 24, March, pp 217-228.<br />

Arts, D. G. T., Keizer, N. F. d. & Scheffer, G.-J. (2002) "Defining and Improving Data Quality in Medical Registries: A<br />

Literature Review, Case Study, and Generic Framework". Journal of the<br />

American Medical Informatics Association, vol. 9, no.6, Nov/Dec, pp. 600-611.<br />

Caballerro, I., Caro, A., Calero, C. & Piattini, M. (2008) "IQM3: Information Quality Management Maturity Model".<br />

Journal of Universal Computer Science, vol. 14, no. 2, December, pp 3658-3685.<br />

Caccia-Bava, M. d. C., Guimaraes, T. & Harrington, S. J. (2006) "Hospital organization culture, capacity to innovate<br />

and success in technology adoption". Journal of Health Organization and Management, vol. 20, no. 3, pp 194-<br />

217.<br />

DeLone, W. H. & McLean, E. R. (1992) "Information Systems Success: The Quest for Dependent Variable".<br />

Information Systems Research, vol. 3, no. 1, pp. 60-95.<br />

English, L. P. (1999) Improving Data Warehouse and Business Information Quality, New York, John Wiley & Sons,<br />

Inc.<br />

Gaikwad, R., Sketris, I., Shepherd, M. & Duffy, J. (2007) "Evaluation of accuracy of drug interaction alerts triggered<br />

by two electronic medical record systems in primary healthcare". Health Informatics Journal, vol. 13, no. 3,<br />

September, pp. 163-177.<br />

Garrib, A., Stoops, N., McKenzie, A., Dlamini, L., Govender, T., Rhode, J. & Herbst, K. (2008) "An evaluation of the<br />

District Health Information System in rural South Africa". S Afr Med, vol. 98, no. 7, July, pp. 549-522.<br />

Greenhalgh, T., Potts, H. W. W., Wong, G., Bark, P. & Swinglehurst, D. (2009) "Tensions and Paradoxes in<br />

Electronic Patient Record Research: A Systematic Literature Review Using the Meta-narrative Method". The<br />

Milbank Quarterly, vol. 87, no. 4, pp 729-788.<br />

Heeks, R. (2006) "Health information systems: Failure, success and improvisation". International Journal of Medical<br />

Informatics, vol. 75, no. 2, pp 125-137.<br />

Institute of Medicine (2000) To Err is Human: Building a Safer Health System, Washington D.C., National Academy<br />

Press.<br />

Irani, Z. (2010) "Investment evaluation within project management: an information systems perspective". Journal of<br />

the Operational Research Society, vol. 61, no. 6, June, pp. 917-928.<br />

Kaplan, B. & Harris-Salamone, K. D. (2009) "Health IT Success and Failure: Recommendations from Literature and<br />

AMIA Workshop". Journal of the American Medical Informatics Association, vol. 16, no. 3, May-June, pp. 291-<br />

299.<br />

Kerr, K. (2006) The institutionalisation of data quality in the New Zealand health sector. PhD Thesis University of<br />

Auckland.<br />

Macinati, M. S. (2008) "The relationship between quality management systems and organizational performance in<br />

the Italian National Health Service". Health Policy, vol. 85, no. 2, February, pp. 228-241.<br />

Miettinen, M. & Korhonen, M. (2008) "Information Quality in Healthcare: Coherence of Data Compared Between<br />

Organization's Electronic Patient Records". CBMS '08 Proceedings of the 2008 21st IEEE International<br />

Symposium on Computer-Based Medical Systems, pp. 488 - 493<br />

Orfanidis, L., Bamidis, P. D. & Eaglestone, B. (2004) "Data quality Issues in electronic health records: An adaptation<br />

framework for the Greek health system". Health Informatics Journal, vol. 10, no. 1, March, pp. 23-26.PMI (2008)<br />

PMBOK: A Guide to the Project Management Body of Knowledge 4th Edition. Project Management Institute<br />

Communications.<br />

Poon, E. G., Wright, A., Simon, S. R., Jenter, C. A., Kaushal, R., Vol, L. A., Cleary, P. D., Singer, J. A., Tumolo, A. Z.<br />

& Bates, D. W. (2010) "Relationship Between Use of Electronic Health Record Features and Health Care<br />

Quality". Med Care, vol 48, no. 3, March, pp 203-209.<br />

Price, R. J. & Shanks, G. (2005) "Empirical Refinement of a Semiotic Information Quality Framework". Proceedings<br />

of the 38th Hawaii International <strong>Conference</strong> on System Sciences, pp. 1-10.<br />

Sutherland, J. M. & Steinum, O. (2009) "Hospital factors associated with clinical data quality". Health Policy, vol. 91,<br />

no. 3, August, pp. 321-326.<br />

537


Siti Asma Mohammed and Maryati Mohd Yusof<br />

Tayi, G. K. & Ballou, D. P. (1998) "Examining Data Quality". Communications of the ACM, vol. 41, no. 2, February,<br />

pp. 54-57.<br />

Tuten, P. M. (2009) A Model for the Evaluation of IS/IT Investments. PhD Thesis Graduate School of Computer and<br />

Information Sciences. Nova Southeastern University.<br />

Wand, Y. & Wang, R. Y. (1996) "Anchoring Data Quality Dimensions in Ontological Foundations". Communications<br />

of the ACM, vol. 39, no. 11, November, pp. 86-95.<br />

Wang, R. Y. & Strong, D. M. (1996) "Beyond Accuracy: What Data Quality Means to Data Consumers". Journal of<br />

Management Information Systems, vol. 12, no. 4, pp. 5-34.<br />

Yusof, M. M., Kuljis, J., Papazafeiropoulou, A. & Stergioulas, L. K. (2008) "An evaluation framework for Health<br />

Information Systems: human, organization and technology-fit factors (HOT-fit)". International Journal of Medical<br />

Informatics, vol. 77, no. 6, pp. 386-398.<br />

Zwikael, O. (2009) "The Relative Importance of the PMBOK Guide's Nine Knowledge Areas During Project<br />

Planning". Project Management Journal, vol. 40, no. 4, December, pp. 94-103.<br />

538


Developing a Change Management Model for Iranian<br />

<strong>Academic</strong> Libraries: A Delphi Study<br />

Maryam Nakhoda 1 and Sirous Alidousti 2<br />

1<br />

University of Tehran, Tehran, Iran<br />

2<br />

Iranian Research Institute for Information Science and Technology (IranDoc),<br />

Tehran, Iran<br />

mnakhoda@ut.ac.ir<br />

alidousti@irandoc.ac.ir<br />

Abstract: Purpose- This paper aims to develop a change management model for Iranian academic libraries. Change<br />

is inevitable in organizations where it is now experienced as a daily phenomenon. <strong>Academic</strong> libraries perform in the<br />

rapidly growing environment of education and research and they need to be active in facing change in order to<br />

realize their mission. To facilitate the planning and implementation of change, several models have been designed in<br />

the past, but these are either too general or have been developed in specific contexts and cannot easily be adopted<br />

in other contexts. In order to overcome the challenges of change and to improve change management, Iranian<br />

academic libraries need their own change model. Design/methodology/approach-The model was designed based on<br />

a qualitative approach and a Delphi study. A panel of twenty-two experts in Management and Library Science and<br />

Information Science contributed to the design and validated the model in three rounds of Delphi. The input of the<br />

Delphi process was the “Information Services Framework for Managing Change” by Penfold and 54 predefined<br />

actions for change extracted from change models related to libraries. The experts were asked to allocate each action<br />

to one (or more) of the stages of the basic model. The panel members could also add their own preferred actions to<br />

the list. The consensus was measured by percentage agreement (60% or greater) for each item. Findings- Thirteen<br />

actions were added by the experts to the initial list. From the resulting 67 actions, only one of them was rejected by<br />

the panel, leaving 66 actions, which were consolidated to 58. These actions were then divided into five stages:<br />

reviewing the current status, analysing and defining library users, planning, implementing the change, and<br />

maintaining the change. After defining the problem, defining change management and describing the respective<br />

model, the methodology of the study and the findings were discussed and finally suggestions for future work were<br />

made.<br />

Keywords: change; change management; change models; academic libraries; Iran<br />

1. Introduction<br />

Modern organizations encounter more extensive changes. Dealing with these changes effectively<br />

requires planning. Change management and its models discuss how to guide an organization during the<br />

change process. Due to their pivotal role in education, libraries and information centres, especially<br />

academic libraries, need strategies for efficient management of change. This paper presents the research<br />

findings to design a change management model for academic libraries, especially in Iran.<br />

2. Statement of the problem<br />

Studies have shown that more than 70% of all efforts for organizational change have been unsuccessful<br />

(By 2005 and Goodrich 2008) mostly because of unfamiliarity with change management (Cicmil 1999 and<br />

Smith 2006). In addition, libraries encounter changes caused by internal and external stimulants (Mark<br />

1997and Pugh 2007). Hirshon (1999) referred to the importance of change management for the<br />

academic libraries worldwide.<br />

Iran is a developing country with an increasing academic growth. Advances in information technologies<br />

and the rise in the number of universities, students, and researchers (Iran’s statistical yearbook 2008)<br />

pave the way of academic improvements in Iran. Such factors create challenges for Iranian academic<br />

libraries which have to be dealt with in order to fulfil their roles. Therefore the need for change in the<br />

academic libraries’ services, spaces, and materials’ formats would appear. According to the importance<br />

of context in developing the managerial processes, the change model of the Iranian academic libraries<br />

must be devised according to their special status and the environment in which they perform their<br />

activities.<br />

The objective of this study is presenting a change model in order to improve the implementation of<br />

changes in Iranian academic libraries as well as increasing the efficiency by reducing the costs and<br />

redundancy.<br />

539


Maryam Nakhoda and Sirous Alidousti<br />

This study contributes to providing a perfect change model for academic libraries, especially academic<br />

libraries of Iran. Also it is of value to academic libraries’ directors to aid them in managing changes in<br />

their libraries.<br />

3. Change management and change models<br />

Change can be defined as kind of state conversion with its tendency being from a relatively undesirable<br />

state towards a more desirable one. Change management and change models are branches of<br />

management studies. Paton, Beranek, and Smith (2008) mentioned the functions of a change<br />

management plan. In change management at least three fields are studied: the current state of the<br />

organization, the state the organization should reach in the future, and finally how to guide this<br />

conversion from the current state to the desired state.<br />

Several studies have been done on change management in academic libraries (Clarke and Morris 1998;<br />

Konings and Dekker 2005; and Ferguson 2007). However, there is little evidence of efforts to apply<br />

existing change models in academic libraries (Sidorko 2008). The first change models followed a threestage<br />

process, including preparation of the organization for change, change implementation and<br />

consolidation of new procedures in the organizational culture. For instance, Lewin’s change model (1974)<br />

consisted of three stages, namely “unfreezing”, “moving”, and “refreezing” (Gilley, Godek, and Gilley<br />

2009). Herein, the term “model” denotes a systematic description of the stages and activities necessary<br />

to carry out a change. The literature review reveals common characteristics among a relatively large<br />

number of change models. Burke (2002) refers to two common functions of change models, namely,<br />

“conversion of organization’s diverse features to simple and manageable parts” and “helping to focus on<br />

more significant parts” in the change process.<br />

4. Research methodology<br />

In the present study, a qualitative approach and the Delphi method in three rounds were applied. The<br />

Delphi method was firstly employed in 1950s in defence sciences research by the RAND Company and<br />

has been used in other fields as well (Green 2000; Ludwig and Starr 2005). Reaching a judgment based<br />

on consensus could be considered as the outcome of a Delphi study (Linstone and Turoff 1975). The<br />

three major characteristics of the Delphi method are “anonymity” of members to each other, “feedback”<br />

and “iteration” (Ambrosiadou and Goulis 1999). The researcher collects the results of the first round of<br />

the study and sends them back again to the panel of experts. After providing the panel of experts with the<br />

results of the first round, each of them expresses his/her opinion once more. This process may continue<br />

for several rounds until consensus is finally reached (Bradley and Stewart 2002).<br />

Among the applications of the Delphi method, which are close to the aim of the present study, design and<br />

validation of models in other fields can be mentioned (Sharma et al. 2008; Minkman et al. 2009; and<br />

Minkman, Ahaus and Huijsman 2009).<br />

4.1 Selection of Delphi panel members<br />

In Delphi studies, the “panel of informed individuals” or “experts” possess knowledge of the subject under<br />

study. In the determination of these experts, purposive sampling is often used (Hasson, Keeney, and<br />

McKenna 2000), in which participants are selected to solve a specific problem using their knowledge.<br />

Concerning the size of the sample, Okoli and Pawlowski (2004) considered ten to twenty qualified<br />

experts as a valid sample.<br />

In the current research, members of the Delphi panel were selected by purposive sampling. A list of<br />

twenty-nine individuals appropriate for participating in this study was initially prepared. Twenty-two of<br />

them expressed their willingness to participate in the first round. The characteristics of the participants<br />

were as follows:<br />

Current director of a central academic library: eight individuals, considering that in Iran most of the<br />

academic (university) libraries are considered “central” academic libraries (Haghighi 2006);<br />

Ex-director of a central academic library: six individuals;<br />

Current assistant of a central academic library: four individuals;<br />

Faculty member of Librarianship and Information Science familiar with the management of academic<br />

libraries or with a background as a manager/consultant/executive responsible for projects for change<br />

in different organizations (such as libraries): one individual;<br />

540


Maryam Nakhoda and Sirous Alidousti<br />

Faculty member of management familiar with the management of academic libraries or with a<br />

background as a manager/consultant/executive responsible for projects for change in different<br />

organizations (such as libraries): three individuals.<br />

Among the current directors of central academic libraries in the panel, seven of them have a PhD degree<br />

in Librarianship and Information Science. One of them has a PhD degree in Mechanical Engineering. The<br />

average duration of their experience as the director of an academic library is 6.5 years. Of the members<br />

with a background as the director of a central academic library, four of them have PhD degrees in<br />

Librarianship and Information Science and two of them have a PhD degree in Management in the Wood<br />

and Paper Sciences industry. The average duration of their experience as the director of a central<br />

academic library is 5.6 years. Of the assistants of central academic libraries who were members of the<br />

panel, three have M.A. degrees in Librarianship and Information Science and one has a M.A. degree in<br />

Management. The average duration of their experience as an assistant in a central academic library is<br />

7.5 years. The faculty member of Librarianship and Information Science, who was member of the panel,<br />

has twenty-five years of experience, and the faculty members of Management in the panel have 19.3<br />

years of experience in this position.<br />

4.2 Preparation of Delphi questionnaire<br />

The design of the Delphi questionnaire was based upon a model that was earlier selected to include two<br />

stages. Of the twenty-nine change models found in the literature for the first stage, three suitable models<br />

for libraries were selected. In the second stage, using consensus method and analysis techniques related<br />

to Multiple Attribute Decision Making (MADM), Penfold’s model (1999), called “information services<br />

framework for managing change” was chosen as the basic model for the Delphi process (Nakhoda,<br />

Fadaie, and Alidousti 2010).<br />

To prepare the Delphi questionnaire, the essential activities for change management in academic<br />

libraries were determined by reviewing appropriate change models to libraries. So, Kotter’s change model<br />

(Kotter 1998), Penfold’s change model (Penfold 1999), and Curzon’s change model known as the “cycle<br />

of change model” (Curzon 2005) were studied to extract the necessary activities to bring about change.<br />

Besides these three models, Lewin’s change model called the “action research model” was also analysed<br />

for the extraction of activities. In comparison with current approaches towards change, this model is<br />

fundamental and among the oldest of the change models (Cummings and Worley 2001). Each of the four<br />

change models mentioned is composed of a series of stages. In turn, each stage of these models<br />

consists of several actions. The resulting collection and summarization of 54 actions from these models<br />

were included in the first round of the Delphi questionnaire as suggested actions for change.<br />

Hence, the main two elements of the first round of the Delphi questionnaire were the five stages of the<br />

basic model (Penfold’s model) and proposed actions for implementing change in academic libraries. The<br />

development and confirmation of the model, based upon conditions in Iranian academic libraries, were<br />

accomplished by members of Delphi panel. The consensus level in this study had a numerical value of<br />

60% of all respondents (Williams and Webb 1994). In other words, in the case that at least 60% of<br />

respondents in second and third rounds were to select a specific stage for a suggested action, this stage<br />

would be considered as one of consensus.<br />

5. Findings and conclusion<br />

In the first round questionnaire, a matrix was provided, composed of five stages of the basic model (in the<br />

first row) and 54 suggested actions for implementing change in academic libraries (in the rightmost<br />

column). To determine its necessity to the change model, each action was first specified by two options,<br />

i.e., yes or no. In the case of a positive response, the respondent was required to select the stage of the<br />

basic model that they considered the most suitable for that action. At the end of questionnaire, the<br />

respondent could write down those suggested actions for implementing change in academic libraries,<br />

which he/she thought, should be added to the list. Among the twenty-two questionnaires sent in the first<br />

round, three of them were not completed, so, nineteen questionnaires (equal to 86.3%) were submitted.<br />

Some Delphi panel members selected more than one stage in the allocation of actions to the Penfold'<br />

model. Furthermore, some members considered some actions not suitable for any of the stages. Table 1<br />

provides the actions added by panel members in the first round.<br />

541


Maryam Nakhoda and Sirous Alidousti<br />

Table 1: Actions added by panel members in the first round<br />

No. Description of actions<br />

1 Preparing an evaluation report of changes<br />

2 Standardizing the services<br />

3 Standardizing the necessary space in the library<br />

4 Reinforcing the expert human resource in the library<br />

5 Convincing top managers of the necessity of change<br />

6 Convincing top managers of the library regarding change and its probable costs<br />

7 Convincing top managers of the library regarding advantages of change<br />

8 Institutionalizing the need for change in the library<br />

9 Forming the culture required for each change<br />

10 Separating plans related to software and hardware changes in the library<br />

Handing out forms to users concerning whether or not to implement a change, accompanied by expressing<br />

11<br />

the positive and negative aspects of changes<br />

Forming a planning group whose members are selected from authorities of the parent organization, library<br />

12<br />

managers, users and staff for comprehensive decision-making<br />

13 Approving the main changes in the university to assure that they will be followed by future managers<br />

In the second round, the responses of each member in first round as well as the frequency of responses<br />

of all Delphi panel members were provided to panel members and they were again asked to express their<br />

opinions. It should be noted that, in this round, the number of suggested actions for implementation of<br />

change was again 54, and an opportunity was again provided to panel members for further judgment,<br />

considering the opinions of the group. The respondents were then asked to review the actions added by<br />

panel members to the list of 54 actions in the first round, and to allocate those actions that they find<br />

necessary to the most suitable stage(s) of Penfold’s model. All respondents in this round had also<br />

participated in the previous round. After collecting the data related to the second round of the Delphi<br />

method and calculating the frequencies of responses, a consensus criterion was applied to the data. The<br />

consensus level was considered as at least 60% of respondents (10 or more members).Therefore,<br />

thirteen actions remained below consensus level.<br />

In third round questionnaire, the responses of each expert in second round and the respective stages<br />

with the highest frequencies were provided. In this stage, frequencies of less than six were not<br />

mentioned. The respondents were asked to first identify the necessity of each of the thirteen actions, and<br />

then allocate each of them to the most appropriate stage(s).Out of the seventeen questionnaires that<br />

were sent in this stage, sixteen of them were filled out and returned. All respondents in this round had<br />

also participated in the previous round. After collecting the data related to the third round of the Delphi<br />

method, the consensus criterion was once again applied to data. The consensus level was set to at least<br />

60% of respondents in this round (nine or more members). Among the thirteen actions that were judged<br />

again, five of them were added to the list of actions by the panel members from the first round. Only the<br />

action, “terminating the contract with an organizational development consultant”, was not identified as<br />

being necessary in a change model for Iranian academic libraries, according to the votes of the majority<br />

of respondents (ten members). In total, 66 actions were allocated to the different stages of the basic<br />

model of the study (Penfold’s change model).<br />

During revision of the Delphi results, it was observed that some actions were covered by some others<br />

and had similar content. Hence, combining these actions into a single action would simplify the model. It<br />

was revealed that eight actions out of the 66 actions had such characteristics. These actions were added<br />

by the panel members in the first round and passed the consensus criterion in the second round.<br />

Ultimately, 58 actions were yielded and each of them belonged to a stage of the basic model. Table 2<br />

provides the final Delphi model, i.e., the stages and actions in the change management model for Iranian<br />

academic libraries. The number of actions does not denote the order of implementing the actions.<br />

Table 2: Change model for Iranian academic libraries<br />

Actions in stage 1: Reviewing the current state<br />

1. Recognising the problem<br />

2. Understanding the nature of change (positive/negative or permanent/temporary)<br />

3. Carrying out SWOT (analysis of strengths, weaknesses, opportunities and threatens) for the library<br />

4. Carrying out PEST (political, environmental, social and technological analysis) for the library<br />

5. Analysing the library’s current plans and strategies<br />

6. Identifying the key factors for success (activities, scope and key conditions) to reach the library’s goals<br />

542


Maryam Nakhoda and Sirous Alidousti<br />

Actions in stage 2: Analysing and identifying the users<br />

1. Evaluating users’ needs and making practical suggestions<br />

2. Identifying library users<br />

3. Handing out forms to users concerning whether or not to implement a change, accompanied by an opportunity<br />

to express positive and negative aspects of changes<br />

Actions in stage 3: Planning<br />

1. Consulting with a behavioural sciences expert regarding the management of the staff’s reactions towards<br />

change<br />

2. Contracting with an organizational development consultant from outside of the library<br />

3. Studying how to obtain and keep the existing information in the library and its costs<br />

4. Studying the vision and mission of the library<br />

5. Studying the general and specific objectives<br />

6. Determining the library services, users of services and requirements for providing services<br />

7. Acquiring information concerning the problem and initial recognition of possible solutions<br />

8. Comparing the library services and processes with similar libraries and the libraries in other sections<br />

9. Providing the acquired information to users or key groups which will be influenced by the change<br />

10. Discussing the pressures and problems with the staff in order to prepare the library for change<br />

11. Holding meetings and discussion with the library staff concerning change<br />

12. Making the decision about whether or not planning is needed for a specific change (evaluating the<br />

advantages, disadvantages and probable impacts of change)<br />

13. Making an agreement on the problem-solving method(s) among the group responsible for change<br />

14. Forming a powerful executive group for planning the change<br />

15. Describing the vision of the change, including its goal and direction<br />

16. Making strategies for reaching the change vision<br />

17. Discussing the change vision with the library staff<br />

18. Searching and finding organizational development resources and support of other sections for implementing<br />

the change<br />

19. Assigning some budget for unpredictable situations in the change process<br />

20. Searching for the source of the library staff’s resistance to change<br />

21. Making the plan for the implementation of change<br />

22. Participating in the strategic planning of the parent organization and developing it<br />

23. Analysing costs and proving the value of the library to the parent organization<br />

24. Separating the plans related to software and hardware changes in library<br />

25. Standardizing the services provided in an academic library<br />

26. Standardizing the necessary space in the academic library<br />

Actions in stage 4: Implementing the change<br />

1. Training in new behaviours which are in accordance with change practically, by the group guiding the change<br />

2. Alleviating barriers to change (personal, structural or those related to systems or skills) in the library<br />

3. Piloting the change in a limited part of library<br />

4. Converting the state of the library from its current state into the desired state<br />

5. Encouraging team work and communication among the library staff<br />

6. Keeping open the communication channels between the director of the library and staff during change<br />

7. Applying the required modifications in the process of execution of the change<br />

8. Analysing of the order of the actions carried out for change<br />

9. Reinforcing the expert human resources in the library<br />

Actions in stage 5: Maintaining change achievements<br />

1. Highlighting advances in the library’s performance (as a result of change)and informing the staff about them<br />

2. Recognizing and appreciating the staff who had a role in the improvements<br />

3. Supervising and evaluating the activities related to change periodically in order to measure their level of<br />

progress<br />

4. Determining the time for evaluating the changes made<br />

5. Determining persons for evaluating the changes made<br />

6. Determining the method(s) for evaluating the changes made<br />

7. Acquiring new information after performing actions, in order to evaluate and determine the impacts of the<br />

actions carried out<br />

8. Disseminating the results of change evaluation among library staff<br />

9. Growing the staff who can assist the realization of the change vision<br />

10. Reinforcing new behaviours caused by change<br />

11. Explaining to the staff the relationship between the new behaviours caused by change and library’s success<br />

12. Making sure that the next top manager will follow the performed change<br />

13. Measuring the satisfaction of the library users<br />

14. Disseminating information about the library’s new services<br />

543


6. Discussion<br />

Maryam Nakhoda and Sirous Alidousti<br />

Approved actions in the model have been confirmed in previous studies too. Recognising the problem<br />

has been considered as an action to recognize issues creating the need for change (Harigopal 2001;<br />

Cummings and Worley 2001). Other actions in stage one will clarify different aspects of change (Penfold<br />

1999). The second stage contains actions to know library users, their needs, and points of view much<br />

better. Adcroft, Willis, and Hurst (2008) pointed to the importance of this knowledge. Considering stage<br />

three, it has been suggested to use the help of consultants to manage psychological and organizational<br />

dimensions in the change process (Lippitt 1972; Harigopal 2001; and Cummings and Worley 2001).<br />

According to Burnes (1992), change could be carried out after the diagnosis of the problem by gathering<br />

relative information. There are actions in planning stage to discuss the change with staff (Leavitt 1965;<br />

Porras and Silvers 1991; and Burke and Litwin 1992). Forming the planning group for change and<br />

assigning budget have been confirmed respectively by Kotter (1998) and Vollmann (1996), Austin and<br />

Bartunek (2006), and Nilakant and Ramnarayan (2006). Software and hardware changes in the library<br />

could be separated in planning stage due to their specifications. In moving from the current stage to the<br />

desired one in stage four (Burnes 1992), developing the library staff, providing them with necessary skills,<br />

and alleviating barriers to change are significant actions (Kotter 1998; Curzon 2005; and Sidorko 2008).<br />

Encouraging team work and communications were confirmed by Kotter (1998), Penfold (1999), and<br />

Curzon (2005) as necessary actions for change. Cummings and Worley (2001) pointed to the evaluation<br />

of change process during the implementation. In stage five, highlighting the successful results of change<br />

and appreciating the helpful staff would be effective in maintaining the change (Kotter 1998). Evaluating<br />

the activities related to change, disseminating the results and information about new services, and<br />

measuring the users’ satisfaction could help the library director in making decision to continue or give up<br />

the change (Penfold 1999). Also in stage five new behaviours caused by change would be reinforced and<br />

the staff would know the relation between their role in change and the library success. The commitment<br />

of next top manager to follow change could help the effective changes to be more permanent (Kotter<br />

1998).<br />

7. Limitations and suggestions<br />

The first limitation of this research is originated form the nature of Delphi method, as the results of Delphi<br />

study could be altered by using a different panel of experts. In addition; the final model of this study<br />

needs to be validated in action to gain the empirical validity too.<br />

In future studies, the design of instruments for the model’s actions can be studied. Regarding the actions<br />

in the proposed change model, another study can be devoted to the order of these actions in the model.<br />

The time needed for performing each action is another issue for future research<br />

The directors of Iranian academic libraries can apply the suggested model to improve the process of<br />

change management in these libraries. The knowledge and skills for managing change could be included<br />

in the lessons of Library and Information science. This would help the students to apply necessary<br />

techniques in their working environments.<br />

References<br />

Adcroft, A., Willis, R., and Hurst, J. (2008) “A new model for managing change: the holistic view”, The Journal of<br />

Business Strategy, Vol. 29, No.1, pp. 40 -57.<br />

Ambrosiadou B. V. and Goulis D. G. (1999) “The DELPHI method as a consensus and knowledge acquisition tool for<br />

the evaluation of the DIABETES system for insulin administration” Medical Informatics, Vol. 24, No. 4, pp. 257-<br />

268.<br />

Austin, J. R, and Bartunek, J. M. (2006) “Theories and Practices of Organizational Development”, Organization<br />

Development: A Jossey-Bass Reader. Editor Gallos, J. V. San Francisco: JOSSEY-BASS, John Wiley & Sons,<br />

Inc.<br />

Bradley, L. and Stewart, K. (2002) “A Delphi study of the drivers and inhibitors of Internet banking”, The International<br />

Journal of Bank Marketing, Vol. 20, No. 6, pp. 250-260.<br />

Burke, W. W. (2002) Organization Change: Theory and Practice, Sage publications, London.<br />

Burke, W.W and Litwin, G. H. (1992) “A Causal Model of Organizational Performance and Change”, Journal of<br />

Management, Vol. 18, No.3, pp. 523-545.<br />

Burnes, B. (1992) Managing Change: a strategic approach to organizational development and renewal, Pitman<br />

Publishing, London.<br />

By, R. T. (2005) “Organizational Change Management: A Critical Review”, Journal of Change Management, Vol. 5,<br />

No. 4, pp. 369-380.<br />

Cicmil, S. (1999) “An insight into management of organizational change projects”, Journal of Workplace Learning,<br />

Vol. 11, No. 1, pp. 5-15.<br />

544


Maryam Nakhoda and Sirous Alidousti<br />

Clarke, L. and Morris, A. (1998) “Library system migration: a case study of change management at Oxford<br />

University”, Libri: International Journal of Libraries & Information Services, Vol.3, No. 48, pp. 153-162.<br />

Cummings, T. G. and Worley, C. G. (2001) Essentials of Organization Development and Change, South-Western<br />

College Publishing, Ohio.<br />

Curzon, S. C. (2005) Managing Change: A How-To-Do-It Manual for Librarians, Revised Edition, Neal-Schuman<br />

Publishers, New York.<br />

Ferguson, A. W. (2007) “Managing change in the Hong Kong library environment”, Library Management, Vol. 28, No.<br />

8 / 9, pp. 524-530.<br />

Gilley, A., Godek, M., and Gilley, J. W. (2009) “Change, Resistance, and the Organizational Immune System”,S.A.M.<br />

Advanced Management Journal, Vol. 74, No. 4, pp. 4-10.<br />

Goodrich, D. F. (2008) “The Relationship of the Theory of Constraints Implementation to Change Management<br />

Integration in Professional Service Organizations” Doctor of Business Administration Diss, Nova Southeastern<br />

University.<br />

Green, J. W. (2000) “Delphi Method in Web Site Selection: Using the Experts”, The Reference Librarian, No. 69/70,<br />

pp. 299-310.<br />

Haghighi, Mahmud(2006) “<strong>Academic</strong> Libraries”, Encyclopaedia of Library and Information Science, By F.<br />

Khosravi.Senior editor: E. Afshar, National Library of Iran, Tehran, pp. 1462-1481.<br />

Harigopal, K. (2001) “Management of Organizational Change: Leveraging Transformation”, Response Books, New<br />

Delhi.<br />

Hasson, F., Keeney, S., and McKenna, H. (2000) “Research Guidelines for the Delphi Survey Technique”, Journal of<br />

Advanced Nursing, Vol. 32, No. 4, pp. 1008-1015.<br />

Hirshon, A. (1999) “Libraries, Consortia, and Change Management”, The Journal of <strong>Academic</strong> Librarianship, Vol. 25,<br />

No. 2, pp. 124-126.<br />

Iran’s Statistical Yearbook of 2007 (2008) [Online].http://salnameh.amar-kr.ir:7001<br />

Konings, k. and Dekker, R. (2005) “Strategy and Change management in Delft University of Technology Library”,<br />

New Review of Information Networking, Vol. 11, No. 1, pp. 109- 121.<br />

Kotter, J. P. (1998) “Leading change: Why transformation efforts fail”. Harvard Business Review on Change, Harvard<br />

Business School Publishing, Boston, pp. 1- 20.<br />

Leavitt, H. J. (1965) “Applying organizational change in industry: Structural, technological, and humanistic<br />

approaches”, Handbook of Organizations, edited by J. M. March, Rand McNally, Chicago, pp. 1144-1170<br />

Linstone, H. A. and Turoff M. (1975) The Delphi Method: Techniques and Applications. Reading, MA: Columbia<br />

University Press.<br />

Lippitt, G. (1972) “Criteria for Selecting, Evaluating and Developing Consultants”, Training and Development Journal,<br />

Vol. 28, pp. 10-15.<br />

Ludwig, L. and Starr, S. (2005) “Library as place: results of a Delphi study”, Journal of Medical Library Association,<br />

Vol. 93, No. 3, pp. 315-326.<br />

Mark, G. (1997) “Reengineering library services for the digital age”, APLIS, Vol. 10, No. 4, pp 22-32.<br />

Minkman, M. et al (2009) “A quality management model for integrated care: results of a Delphi and Concept Mapping<br />

study”, International Journal for Quality in Health Care, Vol. 21, No. 1, pp 66–75.<br />

Minkman, M. MN, Ahaus, K. TB., and Huijsman R. (2009) “A four phase development model for integrated care<br />

services in the Netherlands”, BMC Health Services Research, Vol. 9, No. 42, pp 1-<br />

11.[Online]http://www.biomedcentral.com/1472-6963/9/42<br />

Nakhoda, M., Fadaei, G. R., and Alidousti, S. (2010) “Application of Two Decision Making Methods in Analyzing<br />

Experts’ Ranking of Change Models”,[Online] http://jist.irandoc.ac.ir/browse.php?a_code=A-10-1-<br />

132&sid=1&slc_lang=fa<br />

Nilakant, V. and Ramnarayan, S. (2006) Change Management: Altering Mindsets in a Global Context, Response<br />

Books, New Delhi.<br />

Okoli, C., and Pawlowski, S. D. (2004) “The Delphi method as a research tool: An example, design considerations<br />

and applications”, Information& Management, Vol.42, No. 1, pp. 15-30.<br />

Paton, B., Beranek, L., and Smith, I. (2008) “The transit lounge: a view of organizational change from a point in the<br />

journey”, Library Management, Vol. 29, No. 1 / 2, pp. 87-103.<br />

Penfold, S. (1999) Change Management for Information Services, Bowker-Saur, London.<br />

Porras, J. I. and Silvers, R. C. (1991) “Organization Development and Transformation”, Annual Review of<br />

Psychology, Vol. 42, pp. 51-78.<br />

Pugh, L. (2007) Change Management in Information Services, Ashgate Publishing, Burlington.<br />

Sharma, R. V. et al (2008) “Beyond the digital divide: a conceptual framework for analysing knowledge societies”,<br />

Journal of Knowledge Management, Vol. 12, No. 5, pp. 151-164.<br />

Sidorko, P. E. (2008) “Transforming library and higher education support services: can change models help?”Library<br />

Management, Vol. 29, No. 4 / 5, pp. 307-318.<br />

Smith, I. (2006) “Continuing professional development and workplace learning_15: Achieving successful<br />

organizational change_ do’s and don’ts of change management”, Library Management, Vol. 27, No. 4 / 5, pp.<br />

300- 306.<br />

Vollmann, T. E. (1996) the Transformation Imperative: Achieving Market Dominance through Radical Change,<br />

Harvard Business School Press, Boston.<br />

Williams P. L. and C. Webb (1994) “The Delphi technique: a methodological discussion”, Journal of Advanced<br />

Nursing, No.19, pp. 180-186.<br />

545


Adding Action to the Information Audit<br />

Huan Vo-Tran<br />

RMIT University, Melbourne, Australia<br />

huan.vo-tran@rmit.edu.au<br />

Abstract: The Information Audit (IA) has long been seen as an important tool within the Information Management<br />

field, with its origins stemming from financial audits. It is used extensively in libraries as an improvement tool and,<br />

although many have tried to define it, such as Guy St. Clair (1997), Orna (1999) and Henczel (2001a), there is still no<br />

general consensus on a definition, or the steps taken to achieve it. Whatever form it may take, it is agreed that to<br />

undertake such a task requires a structured approach. The following study will propose a hybrid approach in which<br />

Henczel’s seven-stage Information Audit model will be coupled with the Action Research (AR) methodology in order<br />

to assist a mid-sized architectural practice to manage their information throughout the architectural design process,<br />

and, in particular, as they attempt to design a new academic building for a prominent Australian university.<br />

Keywords: information audit, information management, architectural design process, action research<br />

1. Introduction<br />

As organisations begin to create and accumulate information, they tend to lose sight of how to mange it,<br />

and what information they actually possess. We are currently seeing this information being created at an<br />

astonishing rate, and manifesting itself in many forms. Organisations such as architectural firms are not<br />

only dealing with traditional paper-based documents, but are now also dealing with an abundance of<br />

unstructured data such as emails, pictures from building sites, technical drawings, formal documents<br />

from governing organisations, handwritten notes taken during interviews with clients and even transcripts<br />

collected from focus group activities.<br />

In order for an organisation to achieve high performance it is stated that, “effective information<br />

management is the key” (Roglaski 2006), but the sheer amount of unstructured data being created by<br />

these organisations present it with many challenges. Rogalski continues by stating that, “finding the right<br />

information is difficult, information is not well leveraged among partners and it is not coming together in<br />

ways that will yield useful new insights” (Roglaski 2006).<br />

However, before an organisation can effectively and efficiently manage the information that they create,<br />

possess and disseminate, they must first complete an audit of their current information practices. An<br />

Information Audit (IA) can be seen as a tool that can be used to assist them with this process. Botha and<br />

Boon (2003) suggest that an Information Audit can be defined as:<br />

The systematic examination of the information resources, information use, information flows<br />

and the management of these in an organisation. It involves the identification of users’<br />

information needs and how effectively (or not) these are being met.<br />

The study will attempt to audit the information created by a mid-sized architectural practice as they cycle<br />

through the architectural design process in order to design a new academic building for a prominent<br />

Australian university. This will be done through the use of a hybrid methodological approach in which Sue<br />

Henczel’s seven-stage Information Audit model will be combined with Kemmis and McTaggart’s Action<br />

Research (AR) model.<br />

2. Literature review<br />

2.1 Information Audit (IA)<br />

Over the last three decades there have been many attempts to define what an IA is, and what it should<br />

encompass. Yet to date, there is still no universal consensus. Early IA definitions (Reynolds 1980; Burk &<br />

Horton 1988) tended to focus on more formal information sources with a strong emphasis on document<br />

management, while recent approaches (Buchanan & Gibb 1998; Henczel 2001; Orna 1999) have moved<br />

away from this narrow approach to recognise and incorporate the importance of organisational<br />

approaches and the broad range of information resources.<br />

In order to demonstrate the differences in approaches and definitions of Information Audits a brief<br />

examination of some of the definitions is provided :<br />

546


Huan Vo-Tran<br />

LaRosa (1991)<br />

… A systematic method of exploring and analysing where a library’s various publics are<br />

going strategically, and determining the challenges and obstacles facing those publics. The<br />

audit, which raises questions about where and when these publics find and use information,<br />

gives the library a better understanding of the present and future needs of its constituents<br />

which in turn helps the library determine its most appropriate strategic direction.<br />

St. Clair (1997)<br />

A process that examines how well the organisation’s information needs and deliverables<br />

connects with the organisational missions, needs and goals and objectives.<br />

Buchanan and Gibb (1998)<br />

Discovering, monitoring and evaluating an organisation’s information resources in order to<br />

implement, maintain, or improve the organisation’s management of information.<br />

Orna (1999)<br />

A systematic evaluation of information use, resources and flows, with verification by<br />

reference to both people and existing documents, in order to establish the extent to which<br />

they are contributing to an organisation’s objectives.<br />

Henczel (2001b)<br />

Is a process that will effectively determine the current information environment by identifying<br />

what information is required to meet the needs of the organisation. It establishes what<br />

information is currently supplied, and allows matching of the two to identify gaps,<br />

inconsistencies and duplications. The process will also facilitate the mapping of information<br />

flows throughout the organisation and between the organisation and its external environment<br />

to enable the identification of bottlenecks and inefficiencies.<br />

As we can see from the definitions listed above, a majority focus on the organisations and the<br />

management of their information. It is only LaRosa who makes reference to the end users of this<br />

information, and only a few make reference to the use of Information Technology (IT) in their primary<br />

definition of IA, which can now be seen as an essential part of every organisation and the way they<br />

manage their information.<br />

2.2 Approaches to information audits<br />

Buchanan and Gibb (2007) describe the IA as being, “central to the effective organisational management<br />

of information, however there is evidence from the field that IA is neither fully accepted nor commonly<br />

practiced” (Buchanan & Gibb 2007). However, when an IA is executed by an information practitioner it<br />

can be seen as a very costly exercise for an organisation due to the time and resources that must be<br />

allocated for such an undertaking.<br />

Currently there is no standard or agreed methodological approach within the field, and it is generally left<br />

to the practitioner to sort through a myriad of academic and proprietary methods – some more<br />

comprehensive than others. Once an appropriate methodology has been selected, the practitioner is<br />

required to identify “the numerous tools and technique(s) required to support the methodological process"<br />

(Buchanan & Gibb 2007).<br />

Many may also argue that a standard for IA is not required, as each organisation needs to be treated as<br />

a separate entity and requires a different approach. Buchanan and Gibb (2007) also suggest that there is<br />

a, “lack of an agreed methodological approach” which in turn makes the selection of the methodology<br />

somewhat challenging. To add to the complexity of methodology selection, there has also been limited<br />

empirical evidence regarding the usability of the existing approaches.<br />

In support of Buchanan and Gibb's (2007) argument, Botha and Boon (2003) concluded that, “more<br />

research is required on the topic of information and more of the methodologies need to be tested in<br />

practice” (Botha & Boon 2003), which in turn will allow both practitioners and academics to develop more<br />

reliable IA methodologies that can be confidently used and re-used.<br />

547


2.3 Information audit methodologies<br />

Huan Vo-Tran<br />

As stated in section 2.2 there has been a myriad of methodologies created and adopted. The table below<br />

provides an outline of some of the major methodologies, in chronological order.<br />

Table 1: Outline of some major information audit methodologies in chronological order<br />

Year Author Brief description<br />

1987 Worlock Worlock discusses a framework of headings for the auditing process after testing it<br />

out in various environments and suggests that the judgement of these headings<br />

rests with the person undertaking the audit. In total, there are five headings, each of<br />

which should not be seen as being mutually exclusive. The five headings are:<br />

Utility analysis.<br />

Quality values.<br />

Productivity factors.<br />

Implementation criteria.<br />

Strategic impact statements.<br />

1988 Burk and Horton Burk and Horton were the first to develop InfoMap; it was seen as the first IA<br />

methodology developed for widespread use in the industry. Its focus was to evaluate<br />

the information resources using a four-stage process.<br />

Survey staff using questionnaires or surveys.<br />

Measure the information resources against cost/value.<br />

Analyse resources.<br />

Synthesise the findings and map the strengths and weaknesses of the information<br />

1993 Booth and<br />

Haines<br />

resources against the objectives of the organisation.<br />

Booth and Haines made use of the IA for organisational change and for the<br />

development of a new information policy for a regional health authority in the UK.<br />

Their strategy involved five components, which were:<br />

Identify and review the corporate objective.<br />

Decide what information is needed to meet the corporate requirements.<br />

Conduct an IA through the use of questionnaires and interviews to determine if the<br />

current required information exists within the organisation and if so, how it is<br />

currently being utilised.<br />

Address the identified information gaps and problems where possible.<br />

Develop a comprehensive information management policy for the organisation.<br />

Year Author Brief description<br />

1993 Ellis, Barker,<br />

Potter and<br />

Pridegeon<br />

1994 Webb<br />

(cited in Botha<br />

and Boon 2003)<br />

Ellis, Barker, Potter and Pridegeon acknowledge that there are many different<br />

approaches to information auditing; however, they suggest that, to fulfil its function,<br />

an IA must encompass the following:<br />

Establish what the major goals of the organisation/operation are and what kind of<br />

organizational constraints act upon the operational information systems.<br />

Determine the needs of the users.<br />

Inventory the resources available.<br />

Build up a coherent picture of how the system functions from the information<br />

gathered in the first three stages.<br />

Webb describes the IA according to three distinct stages:<br />

Initial audit (inventory).<br />

Collecting the data.<br />

Data analysis.<br />

It can be seen as an operational advisory audit as it looks at how the IA can be used<br />

to audit the current system and how effectively and efficiently the resources are<br />

being used.<br />

1997 St. Clair St. Clair states that the information audit can be grouped into five main areas which<br />

are:<br />

Getting the ball rolling<br />

Conducting interviews<br />

Organising and conducting interviews<br />

Follow up<br />

1997<br />

1998<br />

2007<br />

2008<br />

Buchanan and<br />

Gibb<br />

Typical pitfalls and mistakes.<br />

Buchanan and Gibb studied a number of IA case studies and developed what they<br />

described as a “universal model” for conducting IAs. Their approach was to create<br />

an IA model that could be used in a number of different environments and for the<br />

purposes of developing an effective information strategy for organisations. The<br />

“universal model” proposed by Buchanan and Gibb consists of five phases, these<br />

being:<br />

Promote<br />

548


Huan Vo-Tran<br />

Identify<br />

Analyse<br />

Account<br />

Synthesise.<br />

2001 Henczel Henczel's work in 2001 leveraged off the strengths of Orna and Buchanan and Gibb<br />

to produce a seven-stage auditing process. Henczel also suggests that the use of<br />

the IA should focus more on the strategic direction of the organisation and that it is<br />

the first step in the development of a knowledge audit or knowledge management<br />

strategy. Henczel's seven stages are:<br />

Planning<br />

Data collection<br />

Data analysis<br />

Data evaluation<br />

Communicating recommendations<br />

Implementing recommendations<br />

The information audit as a continuum.<br />

Year Author Brief description<br />

1990<br />

1999<br />

2004<br />

Orna Orna makes a metaphorical reference to the financial audit in her description of the<br />

IA as an, “authoritative examination of accounts with verification by reference to<br />

witnesses and documents” (Orna, 1990). Orna discusses the scope of the IA in<br />

terms of seven phases which are:<br />

Plan.<br />

Investigate the information available in the organisation.<br />

Identify the resources that are available for making information accessible.<br />

Determine information used to further the purposes of the organisation.<br />

Identify those that are responsible for managing and processing the information,<br />

respectively.<br />

Identify and evaluate the information technology that is used to manage information<br />

resources.<br />

Calculate the cost and determine the value of organisational information resources.<br />

Since then, Orna (1999) has developed an alternative auditing process comprising<br />

ten steps:<br />

Conduct a preliminary review to confirm operational/strategic direction<br />

Gain support/resources from management<br />

Gain commitment from the other stakeholders (staff)<br />

Plan, including the project, team, tools and techniques<br />

Identify the IR, information flow and produce a cost/value assessment<br />

Interpret findings based upon current versus desired state<br />

Produce a report to present findings<br />

Implement recommendations<br />

Monitor effects of change<br />

Repeat the IA.<br />

From the analysis of Table 1, we can clearly see that there are many different approaches to conducting<br />

an IA, ranging from Webb’s (1994) three-step process all the way up to Orna's (1999) ten-step process. It<br />

is also of interest to note that not all IAs focus on the same aspects and contain the same level of<br />

structure and detail.<br />

3. Methodology<br />

As stated in the introduction section, the study will attempt to audit the information created, processed<br />

and disseminated by a mid-sized architectural practice as they cycle through the architectural design<br />

process in order to design a new academic building for a prominent Australian university. Although there<br />

are many different methodologies that could have been selected as outlined in section 2.3, the<br />

researcher has opted for a hybrid approach in which two methodologies will be combined. This particular<br />

study will attempt to combine Henczel’s (2001) seven-stage IA model with Kemmis and McTaggart’s AR<br />

spiral.<br />

3.1 Justification<br />

After careful consideration of the methodologies on offer, the researcher has decided to select Henczel’s<br />

seven-stage IA model over the other competing methodologies as it:<br />

Leverages off the strengths from previous work by Orna and Buchanan and Gibb to produce a<br />

seven-stage auditing process.<br />

549


Huan Vo-Tran<br />

Focuses on a strategic direction for the organisation.<br />

Is similar to the AR cycle – Henczel’s model advocates the IA as a continuum where there is an<br />

establishment of a cyclical process.<br />

Allows for the creation of a knowledge audit or knowledge management strategy.<br />

Is a structured, descriptive and methodical process as opposed to that of Worlock (1987).<br />

Has incorporated IT into its overall process.<br />

Has considered management and operational aspects with the submission of a business case in the<br />

planning stage before proceeding.<br />

In addition to Henczel’s seven-stage IA model, the researcher has also incorporated Kemmis and<br />

McTaggart’s AR spiral. This will be utilised in place of stage six – implementing recommendations. The<br />

reason for this selection is that the researcher believes that, although Henczel’s stage six fits well into the<br />

overall IA process, the use of the Kemmis and McTaggart’s AR spiral will serve as a better diagnostic<br />

and implementation tool as it can be seen as a more strategic and structured approach for implementing<br />

change. The use of an AR methodology should also, “assist in solving an organisational problem, or in<br />

some instances takes a step forward in deepening an organisation’s understanding of themselves”<br />

(Emerald, 2008).<br />

3.2 Participants<br />

The study will make use of a mid-sized architectural and urban design practice as a case study. It will<br />

attempt to audit the information created by a mid-sized architectural practice as they cycle through the<br />

architectural design process in order to design a new academic building for a prominent Australian<br />

university. The practice was established in 1996 and is based in the Central Business District (CBD) of<br />

Melbourne, Australia. It is currently being led by a team of five directors and also encompasses ninety<br />

professionals, who include: architects, technical staff, designers and construction staff.<br />

The practice has been involved with, and completed numerous large-scale commissions for both private<br />

and government sectors, with some of their most recent projects including major commercial buildings,<br />

university education and training facilities, and government and defence projects.<br />

3.3 Henczel’s seven-stage IA model<br />

Similar to Orna’s (1990) IA model, Henczel’s IA model also consists of seven main stages. As<br />

demonstrated in Figure 1 each stage is clearlydefined and each stage needs to be complete before the<br />

next one can commence. Listed below is a short description of what is involved in each stage.<br />

Stage one: Planning<br />

Involves planning and the submission of a business case for approval by the organisation before<br />

proceeding on to the data collection.<br />

Stage two: Data Collection<br />

Involves the collection and development of an Information Resource (IR) database and its population<br />

through survey techniques such as questionnaires, personal interviews or focus groups.<br />

Stage three: Data Analysis<br />

Once the data has been collected, it must be organised in a way that allows it to be analysed. It is up to<br />

the person leading the IA to select the most appropriate method for the analysis of the data. Henczel<br />

suggests that, as specialist skills are required to accomplish this, it might be worthwhile to contract<br />

experts outside your organisation.<br />

Stage four: Data Evaluation<br />

At the data evaluation stage, the data begins to show the person conducting the audit a “snapshot” of the<br />

organisation’s information environment. This will facilitate the interpretation and formulation of the<br />

recommendations<br />

550


Huan Vo-Tran<br />

Figure 1: Henczel’s seven-stage information audit model<br />

Stage five: Communicating Recommendations<br />

Henczel suggests that there are many ways of doing this, depending on the size of the organisation and<br />

the scope of the recommendations. However, the key people who need to be kept informed through this<br />

process may include:<br />

Anyone who has championed the audit<br />

Any sponsors<br />

People who directly participated<br />

Those who will be affected by the recommended changes. (Henczel 2001).<br />

The communication of these recommendations may be delivered in many forms and could include:<br />

reports, website or intranet, verbal presentations, seminars and personal feedback to participants.<br />

Stage six: Implementing Recommendations<br />

This stage can be seen as the second-last stage in Henczel’s seven-stage IA model. It involves the<br />

development of a plan for when and how the recommendations outlined in the previous stage will be<br />

implemented. Henczel suggests that the organisation, together with the nature of the recommendations,<br />

will influence how the implementation occurs.<br />

Stage seven: The Information Audit as a Continuum<br />

Henczel suggests that when you reach the end of the audit, it is really only the beginning and that<br />

organisations need to be thinking about the IA as a continuous process to ensure that data gathered<br />

initially can be re-assessed and updated.<br />

3.4 Kemmis and McTaggart’s Action Research spiral<br />

McKay and Marshall (2001) describe how Action Research (AR) in its simplest form involves both action<br />

and participation within a particular field. Its focus is to problem-solve in order to improve the way<br />

processes are performed and services are delivered.<br />

551


Huan Vo-Tran<br />

AR typically makes use of four main phases of continuous change, which are: Plan, Act, Observe and<br />

Reflect<br />

For the purpose of this study, Kemmis and McTaggart’s AR Spiral was selected and will be embedded<br />

within the implementation recommendations stage (stage six) of Henczel’s seven-stage IA model. The<br />

spiral consists of the four phases of AR and is to be carried out collaboratively with the organisation that<br />

is being studied.<br />

Figure 2 represents Kemmis and McTaggart’s AR Spiral (1988). It includes the four main phases of<br />

continuous change (plan, act, observe and reflect). Provided below is a brief description of what happens<br />

within each phase.<br />

Plan:<br />

Develop a plan of critically informed action to improve what is already happening within the organisation<br />

that is being studied.<br />

Act:<br />

Based on the plan formulated, act upon the plan and implement it.<br />

Observe:<br />

Observe the effects of the critically informed action in the context in which it occurs.<br />

Reflect:<br />

Reflect on these effects as the basis for future planning, subsequent critically informed action and so on,<br />

through a cycle of succession cycles. (Kemmis & McTaggart, 1988)<br />

Figure 2: Kemmis and McTaggart’s (1998) action research spiral<br />

3.5 Combination of the two methods<br />

As stated in section 3.1, the study will take on a hybrid approach where Henczel’s seven-stage IA model<br />

and Kemmis and McTaggart’s AR Spiral will be used in conjunction with one another. The study will cycle<br />

through each one of the seven stages of Henczel’s IA model; however, there will be a modification made<br />

552


Huan Vo-Tran<br />

to the sixth stage (implementing recommendations) of the audit instead of making use of Henczel’s sixth<br />

stage, where the outputs should include:<br />

An information policy<br />

Post-implementation strategy<br />

Revised business plan and information management process<br />

An updated information resources database.<br />

With the addition of Kemmis and McTaggart’s AR Spiral, the findings of the previous stages of Henczel’s<br />

seven-stage IA model will inform the planning phase and, in the last component of reflection, will lead into<br />

the final stage, which is the information audit as a continuum. Figure 3 demonstrates how this will be<br />

accomplished.<br />

To what extent is the current<br />

process working and how<br />

can it be streamlined?<br />

Record the process taken<br />

as new information comes<br />

in. Also record the interview<br />

responses from the<br />

architects to see how they<br />

manage and disseminate<br />

this data.<br />

Has this process been<br />

streamlined, if so how? How<br />

important is the collected<br />

data for the next phase in<br />

the architectural design<br />

process?<br />

Follow up on the<br />

observations by interviewing<br />

the architects to see if the<br />

suggested changes have<br />

been beneficial.<br />

In the feasibility stage of the<br />

architectural design<br />

process, information is<br />

collected from a wide range<br />

of sources including focus<br />

group interviews and<br />

emails. This is then placed<br />

in a folder called ‘briefing<br />

folder’. From an information<br />

management point of view,<br />

how can this process be<br />

improved? Is teaching<br />

generic information<br />

management skills to<br />

architects going to improve<br />

the process?<br />

Shift the way architects<br />

think about information<br />

management by suggesting<br />

‘best practice’ tools and<br />

techniques then allowing<br />

them to explore these.<br />

Try questioning the<br />

architects so that the can<br />

critically reflect on their<br />

information management<br />

approaches and see if they<br />

can come up with other<br />

solutions.<br />

Continue general aim but try<br />

and emphasize ‘best<br />

practices’ in information<br />

management.<br />

Record any changes in the<br />

architects’ approach to<br />

information management<br />

after suggestions have been<br />

made.<br />

Figure 3: The incorporation of Kemmis and McTaggart’s action research spiral within stage six of<br />

Henczel’s seven-stage information audit model. (Vo-Tran 2010)<br />

In terms of the techniques that will be used within the combined methodologies, the researcher has<br />

decided to select two commonly used techniques, these being observations and interviews. The use of<br />

observations for this study will allow the researcher to understand, without interfering with the current<br />

practices within the organisation during their the planning phases. Observations can be used again to<br />

see if the changes that have been made during the implementation stages have been beneficial. In<br />

addition to the observations, interviews will be conducted to explore the attitudes and responses to the<br />

changes that have been made. This will also allow the researcher to gather qualitative information in the<br />

form of narratives. It is anticipated that the interviews will take place at the organisation and take on a<br />

planned but unstructured approach.<br />

553


4. Preliminary findings<br />

Huan Vo-Tran<br />

As the study is currently in progress, this section will outline the major findings of the first two stages to<br />

date. The findings will be broken down and reported in terms of Henczel’s seven-stage IA model.<br />

Stage one: planning<br />

A formal letter was submitted to the directors of the architectural practice seeking their permission to<br />

participate in the study. Written permission was granted by the practice and the researcher then sought<br />

and received ethics clearance from their home university.<br />

Adhering to Henczel’s seven-stage IA model, the following tasks and activities were considered and<br />

implemented:<br />

Table 2: Outline of the tasks and activities that were considered and implemented<br />

Task Activity<br />

Developing clear objectives An understanding of what the IA is trying to achieve for both the researcher and<br />

architectural practice<br />

Understanding the organisation and its core business<br />

Identification of the key stakeholders<br />

Defining and scoping the<br />

resource allocation<br />

Physical and information scope<br />

Human, financial, technical and physical resource allocation<br />

Insource, outsource options<br />

Selection of a methodology Data collection, analysis and evaluation<br />

The presentation of the findings and recommendations<br />

Action plan for implementing the recommendations<br />

Development of a<br />

communication strategy<br />

(Henczel 2001a)<br />

Before the audit<br />

During the audit<br />

After the audit<br />

From the tasks and activities listed in the table above, a business case was formulated and presented to<br />

the directors of the architectural practice; this has subsequently been approved and signed off.<br />

Stage two: data collection<br />

Upon completion of the business case, data needed to be collected. This was undertaken using a<br />

combination of techniques, such as onsite interviews and observations. From this, an Information<br />

Resources (IR) database was established and greater understanding of the architectural practice was<br />

achieved. Listed below are some of the findings that resulted from the interviews and observations.<br />

Projects are won through the tendering process, in which the architectural practice presents their<br />

ideas to prospective clients.<br />

Staff at the architectural practice work around project-based teams. They can be working on multiple<br />

projects at any given time depending on workload and expertise.<br />

Each project has one of the architectural practice’s five directors in charge, and these directors may<br />

be spread across no more than three projects.<br />

All project teams are multidiscipline, and may include architects, draftsmen, consultants, engineers<br />

and designers.<br />

Projects may last anywhere between six months and four years, depending on the size and<br />

complexity.<br />

In addition to the practices and processes listed above, the data collection also identified some<br />

challenges in information management the architectural practice is facing:<br />

Storage of documents and images. Documents and images are currently being duplicated in both<br />

physical and electronic formats. These may not be identical as it might take time to update the<br />

documents to reflect either format. In addition to this, electronic copies of documents are stored<br />

within folders according to the file type, e.g. all the PDF files in one location while all the word<br />

documents are in another.<br />

554


Huan Vo-Tran<br />

Recording decisions. The architects have stated that this is one of their biggest problems; any<br />

changes that have been made to a building go straight onto the drawing and there is no record to<br />

why this decision has been made. This causes problems later on when they have no idea why these<br />

changes have been made.<br />

Re-use of information. Although no two projects will be ever the same, there are some elements or<br />

aspects that could be reused. Instead, the architects and draughtsmen spend lengthy amounts of<br />

time trying to re-create almost identical drawings when they could be spending the time working on<br />

other elements of the building.<br />

Transfer of project details to new staff members. As the practices’ staff members move from one<br />

project to another, it takes them time to read through the brief, and attempt to get up to speed as<br />

soon as possible. This is done through reading both the physical and electronic documents, which<br />

may not always contain the most recent changes. This in turn, means that they will have to spend<br />

more time with another staff member going through the finer details.<br />

5. Directions for future research<br />

As the research is currently in progress, it is anticipated that over the next six months the study will move<br />

from the data collection stage into the data analysis and evaluation stages. The data collected so far will<br />

now inform what needs to be done in terms of implementing the recommendations and structuring the<br />

action research component of the hybrid methodology. It is also anticipated that once the IA has been<br />

completed, it will add to the body of knowledge about which Botha and Boon (2003) concluded in a<br />

paper, “more research is required on the topic of information and more of the methodologies need to be<br />

tested in practice” (Botha & Boon 2003).<br />

References<br />

Booth, A. and Haines, M. (1993) "Information audit: whose line is it anyway?" Health Libraries Review, Vol 10, pp<br />

224–232.<br />

Botha, H. and Boon, J. A. (2003) "The Information Audit: Principles and Guidelines", Libri, Vol 53, pp 23–38.<br />

Buchanan, S. and F. Gibb (1998). "The Information Audit: An Integrated Strategic Approach." International Journal of<br />

Information Management, Vol 18, No. 1, pp 29 - 47.<br />

Buchanan, S. and F. Gibb (2007). "The Information Audit: Role and Scope." International Journal of Information<br />

Management, Vol 27, No. 3, pp 159 - 172.<br />

Buchanan, S. and F. Gibb (2008). "The information audit: Methodology selection." International Journal of<br />

Information Management, Vol 28, No. 1, pp 3 - 11.<br />

Burk, C. F. and F. W. Horton (1988). InfoMap: A complete guide to discovering corporate information resources.<br />

Englewood Cliffs, NJ, Prentice-Hall.<br />

Ellis, D., R. Barker, et al. (1993). "Information Audits, Communication Audits and Information Mapping: A Review and<br />

Survey." International Journal of Information Management, Vol 13, No. 2, pp 134 - 151.<br />

Emerald. (2008). "How to... carry out action research." [Online], Emerald,<br />

http://www.emeraldinsight.com/research/guides/methods/action_research.htm.<br />

Orna, E. (1990). Practical information policies: how to manage information flow in organizations. Aldershot, Gower<br />

Publishing Limited.<br />

Orna, E. (1999). Practical Information Policies. Hampshire, Aldershot.<br />

Orna, E. (2004). Information strategy in practice. Aldershot, Gower Publishing Limited.<br />

Dubois, C. P. R. (1995). "The information audit: its contribution to decision making." Library Management, Vol 16, No.<br />

7, pp 20 - 24.<br />

Henczel, S. (2001a). The Information Audit: A Practical Guide. München, K.G.Saur.<br />

Henczel, S. (2001b). "The Information Audit as a First Step Towards Effective Knowledge Management." Information<br />

Outlook, Vol 5, No. 6, pp 13.<br />

Kemmis, S. and R. McTaggart (1988). The Action research planner. Waurn Ponds, Deakin University Press.<br />

McKay, J. and P. Marshall (2001). "The dual imperatives of action research." Information Technology & People, Vol<br />

14, No. 1, pp 46 - 59.<br />

LaRosa, S. (1991). "The corporate information audit." Library Management Quarterly, Vol 14, No. 2, pp 7-9.<br />

Roglaski, S. (2006). "The Rising Importance of Enterprise Content Management." DM Review, Vol 16, No. 11, pp 36.<br />

St. Clair, G. (1997). "The information audit I: defining the process." InfoManage, Vol 4, No. 6, pp 5-6.<br />

Swash, G. D. (1997). "The information audit." Journal of Managerial Psychology, Vol 12, No. 5, pp 312-318.<br />

Vo-Tran, H. (2010). "Auditing the architectural design process", Paper read at 1 st International <strong>Conference</strong> on<br />

Information Management and Evaluation. Cape Town, South Africa.<br />

Worlock, D. (1987). Implementing the information audit. 59th Aslib Annual <strong>Conference</strong>, 'The information Manager'.<br />

University of Sussex, Electronic Publishing Services Ltd. 39: 255-260.<br />

555


Business Intelligence Best Practices for Success<br />

Joseph Woodside<br />

Cleveland State University, USA<br />

joseph.m.woodside@gmail.com<br />

Abstract: Business Intelligence (BI) is in a high adoption and high growth area, as users quickly value the<br />

capabilities and increasingly demand more features to compete in today’s economic climate. However, from a return<br />

on investment (ROI) standpoint, BI is similar to ERP and CRM, in that it has a poor risk/reward profile, as it regularly<br />

runs into cost overruns, due to scope creep and limitless requests for support from end-users (Bernard 2009). Unlike<br />

operational systems which often have specific requirements and implementation completion timelines, BI<br />

environments are constantly evolving to meet business and information requirements (Moss 2007). Given the<br />

complexity of most system implementations, no single measure exists for Business Intelligence success. In order to<br />

effectively evaluate BI success, measures are developed to identify critical implementation factors based on the<br />

research objectives and investigation (Wixom 2001). As an organization progresses in BI maturity, the value of its<br />

activities expands. Successful organizations increasingly utilize analytical approaches to identify and enact modest<br />

improvements that increase profitability and return on business intelligence investments. This paper presents several<br />

key findings, lessons learned, success evaluation methods, and best practices as identified through prior literature<br />

review and a formal empirical study, which extends and enhances prior literature and understanding of BI.<br />

Keywords: BI, business intelligence, critical success factors, implementation, success<br />

1. Road to Business Intelligence<br />

Beginning in the late 1960s, experiments began with Decision Support Systems (DSS), utilizing<br />

computers to analyze data and offer decision-making support. DSS were typically used for narrowly<br />

focused activities such as production planning, investment management, and transportation applications,<br />

and several inputs are required to prepare the analysis (Leidner 1993). With the introduction of software<br />

applications such as SAS and SPSS in the 1970s, statistical software became more available and<br />

accessible to end users. Despite this introduction, DSS did not prosper and evolved into Executive<br />

Support Systems (ESS). (Davenport 2007; Ranjan 2008).<br />

ESS were utilized by executives for viewing firm performance and focused less on decision making<br />

support. The feature found in most ESS was single database access with current organizational<br />

information, in an easy to access manner. ESS usage was also found to be positively related to problem<br />

identification, decision making, and analysis. Other ESS features included non-keyboard interface,<br />

organizational database, drill-down capabilities, trend analysis, exception reports, graphics, and critical<br />

information monitoring. The focus of an ESS was on the organizations day-to-day activities as well as<br />

marketplace indicators. A DSS by contrast was intended to allow on demand decisions and routine<br />

analysis (Leidner 1993). ESS were referred to as high-risk/high-return systems, as the systems serve<br />

executives whose information needs are complex, but also have greater influence. ESS provided<br />

executives easy to use information that supports their critical success objectives. ESS also failed to enjoy<br />

widespread usage due to resistance by executives to hands-on usage (Rainer 1995).<br />

Firms have since made major investments in such systems as enterprise resource planning (ERP),<br />

supply chain management (SCM) and customer relationship management (CRM), yet struggle to achieve<br />

competitive advantage. Firms need streamlined access and analysis of the underlying information in<br />

order to make operational decisions. Strategic organizations sought to improve efficiency through faster<br />

and better-informed decision making, and looked to technology to enhance strategic and tactical results<br />

to improve time to market, connectivity, integration, and visibility into their business. Unbeknownst to<br />

users at the time, this data from these systems was a significant organizational asset, which would later<br />

be leveraged for success and competitive advantage. In order to realize these benefits however, the data<br />

must be developed into an enterprise wide unified view. Construction and integration of knowledge is a<br />

key to succeeding in the competitive global market. Information Technology moved to support day-to-day<br />

operations and all aspects of decision making, with differentiation through technology becoming<br />

increasingly important. New generations of technology savvy users and executives were finding ways to<br />

utilize previously untapped information. The entire field emerged as Business Intelligence (BI) and<br />

includes collection, management, and reporting of decision making data and information. BI capabilities<br />

have consistently been identified as the number one technology priority for organizations according to<br />

current industry surveys (Davenport 2007; Ranjan 2008).<br />

556


2. Business Intelligence<br />

Joseph Woodside<br />

BI is not a single product, application, program, user, area, or system, rather an architecture of integrated<br />

systems that provide users with easy access to and storage of information for decision making and<br />

learning. Competitive pressures cause organizations to continually improve and adapt in order to be<br />

successful in the ever changing business environment, and information is required by employees<br />

throughout all levels of the organization for ongoing decision making (Ranjan 2008; Saha 2007).<br />

Business Intelligence refers to applications and technologies used to gather, capture, access,<br />

consolidate, and analyze information to improve decision making by various horizontal and vertical levels<br />

of stakeholders. These systems capture important metrics on business operations, as well as providing a<br />

mechanism for improved decision making. At the various levels these information items may include<br />

documents, calendars, wikis, links, reports, dashboards, scorecards, search, databases, lists, user<br />

knowledge, and much more. For example, these technologies can help coordinate projects, calendars,<br />

schedules, discuss ideas, review documents, share information, keep in touch with others, utilize Key<br />

Performance Indicators (KPI) to gauge operational status, and generate reporting information ondemand.<br />

The BI process is one that allows large amounts of disparate data to come together into a single<br />

repository and turn that data into meaningful information for decision support processes. BI can include<br />

various forms of analysis, data mining, scorecards, dashboards, metrics, reporting, portals, data<br />

warehouse, OLAP, decision support, knowledge management, etc. This information is available to all<br />

levels of the organization and associated stakeholders, on-demand, and in an easy-to-use fashion (Moss<br />

2007; Ranjan 2008; Saha 2007).<br />

3. Research model<br />

During interviews with executives, managers, and professional staff, before, during, and after BI<br />

implementation and from review of prior literature, a path-analytic BI implementation success model was<br />

proposed as shown in Figure 1. Literature review from information systems was utilized to develop the<br />

research model and relationships between those areas. Prior literature has studied information success<br />

using multiple methods. Researchers are instructed to utilize appropriate success measures based on<br />

the objectives of the study and investigation (Wixom 2001).<br />

Figure 1: BI implementation success model<br />

557


3.1 Collaborative culture<br />

Joseph Woodside<br />

Team work is an important aspect throughout any implementation project. This requires close<br />

cooperation between all department areas and business and technical teams, including top management,<br />

consultants, end-users, and vendors (Bhatti 2005). Aligning resources with business strategies is also<br />

important, as typically there is limited alignment between BI strategy and business strategy (Williams<br />

2007). Organizational learning through shared visions and commitment to learning is an important aspect<br />

to collaborative culture and can be achieved through team work and alignment between units in support<br />

of a common objective and set of goals.<br />

H1: There is a positive relationship between collaborative culture and implementation success<br />

3.2 Customization<br />

Business Process Re-engineering (BPR) involves rethinking and redesigning business processes to<br />

improve key performance measures such as cost and quality of service. Most organizations are required<br />

to modify their existing business processes to fit the application software as a way to limit customizations<br />

(Bhatti 2005). Arnott describes this similarly as the degree of fit between the organization and the<br />

software and hardware (Arnott 2008). Many times an implementation may not meet expectations, due to<br />

an underestimation of change management complexities and encountering resistance to change (Bhatti<br />

2005). In many cases, end users must be trained in the new paradigm after and during BI<br />

implementation. Users must also be trained in understanding and adjusting to changes in business<br />

processes (Williams 2007). BI platforms should be selected which allow for customizations to occur<br />

dynamically through the user interface, this resolves prior system issues which occurred at time of<br />

upgrade of customized applications or modules. The flexibility of most BI platforms allow for<br />

customizations which previously were unavailable in the scope of an enterprise application.<br />

H2: There is a positive relationship between customization and implementation success<br />

3.3 Communication<br />

Wide information sharing and understanding must occur by all stakeholders throughout the<br />

implementation stages and beyond. Communication should start as early as possible to gain<br />

organizational understanding and acceptance (Bhatti 2005). Beyond development of BI and user training,<br />

the vision must be marketed and communicated. The BI applications must be viewed as mission critical<br />

and all users share that vision (Williams 2007). Communication should be started early in the form of<br />

several announcements and organizational newsletters, including email, meeting, and Intranet<br />

announcements. Kick-off meetings with key personnel and staff resources should also take place, along<br />

with regularly recurring meetings.<br />

H3: There is a positive relationship between communication and implementation success<br />

3.4 Project management<br />

Organizations should use structured and formal approach for BI projects. Many projects fail to adequately<br />

account for organizational requirements, resources, and funding necessary to support a successful BI<br />

implementation (Williams 2007). PM includes coordinating, scheduling, scope, and monitoring activities<br />

and resources in line with the project objectives. PM is also responsible for the overall implementation<br />

process and developing organizational support. The DW/BI systems should be developed iteratively<br />

building to a complete application set. (Arnott 2008; Bhatti 2005). Agile methodology was adopted, having<br />

a formal project methodology and a formal project management office for oversight and project tracking is<br />

critical to the project’s success. It is important to establish critical success indicators and metrics from<br />

project inception, to ensure expectations are met and exceeded. To allow improved deployment speed,<br />

an iterative methodology with rapid prototyping should be employed. Parallel user sub-groups should be<br />

established to allow continuous feedback from rapid prototyping and to reduce periods of inactivity.<br />

H4: There is a positive relationship between project management and implementation success<br />

3.5 Resources<br />

Resources can include financial, people, hardware, software and time for project completion. It is also<br />

important to fund new activities required as a result of BI implementation such as meta data<br />

558


Joseph Woodside<br />

management. Resource issues often have a negative impact on implementation success (Arnott 2008;<br />

Williams 2007; Wixom 2001). Consultants are often required due to a knowledge gap and complexity of<br />

new systems. User involvement can occur through requirements gathering, implementation participation,<br />

and use after go-live (Bhatti 2005). Dedicated resources should be assigned to avoid inevitable<br />

competing projects and priorities. Dedicated consultant resources should be allocated to improve timeline<br />

and knowledge transfer for new technology areas. Though consultants should only be utilized as a<br />

temporary solution, as knowledge-loss occurs with continued usage. Dedicated department or area<br />

based resources should be assigned for local subject matter experts and knowledge diffusion.<br />

H5: There is a positive relationship between resources and implementation success<br />

3.6 Top management support<br />

Top management provides the required resources in a direct or indirect manner through financing, as<br />

well as the power and support. Top management is also responsible for setting a clear direction, overall<br />

project objectives, project guidance, representation, and establishing these throughout the organization<br />

(Arnott 2008; Bhatti 2005). Sponsorship across the entire management team, allows others in the<br />

organization to support the project through reducing political resistance, and facilitate participation. Top<br />

management support must include top management champions and are viewed similarly (Wixom 2001).<br />

All top management should be advised of the project by the sponsors, and any issues or concerns<br />

addressed initially. It is vital for the sponsors to continually update top management early in the project,<br />

and as components are released to end-users.<br />

H6: There is a positive relationship between top management support and implementation success<br />

3.7 Training<br />

Training end-users is important to improve knowledge and appropriate use of the system. General BI<br />

concepts, components, demonstration, and use are key training areas. Training should also include<br />

process changes and overall flow of information and integration (Bhatti 2005). Training also includes the<br />

standards and policies that must be followed for the new BI applications, to optimize use of BI by endusers<br />

(Williams 2007). Training modules and materials should be developed prior to the initial go-live,<br />

along governance plans, such as best practices, content and technical standards, and policies and<br />

procedures should be developed for training purposes. In addition, proactive monitoring of training issues<br />

should be immediately addressed to avoid long-term paradigm creation in early stages. New users<br />

should be required to take established training and existing users on an annual basis and directed to the<br />

training materials as questions arise.<br />

H7: There is a positive relationship between training and implementation success<br />

3.8 Vertical integration<br />

In prior BI implementations multiple vendor solutions required purchase, as one vendor did not provide a<br />

fully integrated solution, also other companies chose a best of breed approach based on vendor<br />

offerings. Today, several vendors offer completely integrated solutions with equitable offerings across<br />

services. In one survey organizations utilized an average of 3.2 vendors with 8-13 tools (Howson 2008).<br />

For the small-medium business scenario, utilizing a vertical architecture with a single vendor was<br />

identified as a critical success factor. Due to resource and funding requirements, the ability to match<br />

expert skillsets with a best of breed approach is not possible. The use of a single vendor also improves<br />

delivery time through ease of installation and avoids integration issues that commonly arise when utilizing<br />

multiple vendor solutions vendors. Here, a vertical architecture is defined as having a single vendor<br />

designed and developed BI platform, which includes Knowledge Management, Content Management,<br />

Performance Management, End-User Tools, Querying / Reporting, Analysis, and Database Management<br />

System.<br />

H8: There is a positive relationship between vertical architecture and implementation success<br />

3.9 Success factors<br />

BI project implementation success is measured through perceived success, whether the project<br />

completed timely, completed on budget, and overall satisfaction with the BI (Howson 2008; Wixom 2001).<br />

Table 1 displays the summary of supporting works for construct and hypothesis development.<br />

559


Table 1: Summary of supporting works<br />

Wixom and<br />

Watson (2005)<br />

Joseph Woodside<br />

Bhatti<br />

(2005)<br />

Williams and<br />

Williams (2007)<br />

Howson<br />

(2008)<br />

Implementation Factors<br />

Collaborative Culture X X<br />

Customization X X X<br />

Communication X X<br />

Project Management X X X X<br />

Resources X X X X<br />

Top Management Support X X X<br />

Training X X<br />

Vertical Integration X<br />

Success Factors<br />

Perceived Success X X<br />

Timely Implementation X X<br />

Satisfaction X X<br />

4. Research methodology<br />

Arnott<br />

(2008)<br />

A survey was development followed Moore and Benbasat's (1991) identified stages of item creation,<br />

scale development, and testing: Item creation, in which existing items were utilized from prior literature,<br />

then additional items added to those components which fit the definitions. Scale development, where<br />

similar categories of items were created and refined as needed. Testing, in which sample surveys were<br />

conducted, and then was followed by revisions and larger distribution (Moore 1991). The final survey was<br />

randomized to reduce order effects. The survey uses a seven point likert scale, and measures the level of<br />

agreement with each statement, with 1 being strong disagreement and 7 strong agreement. The survey<br />

was administered to a national healthcare organization who had recently completed a BI implementation.<br />

The survey was reviewed by subject matter experts in and across organizational levels as part of a pretest<br />

to resolve any concerns or address any identified discrepancies, as well as re-word identified items,<br />

or remove non-weighting or duplicative manifest variables. A total of 141 responses were received with<br />

105 (75%) non-managerial users, and 36 (25%) managerial users. Multiple geographic locations were<br />

included, along with multiple staff functions and department areas. Overall 148 responses were received,<br />

with 7 responses removed that were incomplete past general user information, for a total of 141 usable<br />

responses. 106 respondents were non-managerial users, with the remaining 35 respondents holding a<br />

supervisory position through executive position. Users from multiple geographic office locations and<br />

functional areas were represented within the study group, as well as users with short and long-term<br />

tenure at the firm.<br />

5. Data analysis<br />

A path analysis is employed using SmartPLS 2.0 software to analyze the results and determine model fit<br />

(Ringle 2005). A model with significant loadings is developed. Significance of relationships was<br />

determined between implementation factors and success factors. The final model specification includes<br />

the supported paths, all paths were supported for significance at p = 0.01. Results showed that 73.8% of<br />

the variability in implementation success is explained from the model.<br />

Table 2 displays the quality of the model. The average variance extracted is indicated by the column<br />

AVE, and is the average communality for the latent factors in the model. AVE is utilized for convergent<br />

validity, and should be greater than or equal to 0.5, which the latent factors exceed. Composite reliability<br />

is also utilized as Cronbach’s alpha commonly underestimates or overestimates reliability. Composite<br />

reliability follows similarly to Cronbach’s alpha, with 0.80 to be considered good, 0.70 to be considered<br />

acceptable, and 0.60 to be considered for exploratory requirements. Composite reliability exceed 0.90 for<br />

this model. R-Square displays the effect size measure, and is not shown for exogenous constructs. An R-<br />

Square of 0.67 is considered substantial, 0.33 considered moderate, and 0.19 considered weak. The R-<br />

Square for this model is 0.73. Cronbach’s alpha should be equal to at least 0.80 to be considered good,<br />

0.70 to be considered acceptable, and 0.60 to be considered for exploratory requirements. For short<br />

scales, Cronbach’s alpha may be biased, this study utilized 7 point scales for all questions measured,<br />

and Cronbach's alpha exceed 0.80 (Garson 2010; Ringle 2005). Table 3 displays the hypothesis results.<br />

560


Table 2: Measurement model quality<br />

Joseph Woodside<br />

AVE Composite Reliability R Square Cronbachs Alpha<br />

Implementation Factors 0.57986 0.90581 0.878317<br />

Success Factors 0.66762 0.922954 0.738381 0.899059<br />

Table 3: Hypothesis results<br />

Hypothesis Result<br />

H1: There is a positive relationship between collaborative culture and implementation success Supported<br />

H2: There is a positive relationship between customization and implementation success Supported<br />

H3: There is a positive relationship between communication and implementation success Supported<br />

H4: There is a positive relationship between project management and implementation success Supported<br />

H5: There is a positive relationship between resources and implementation success Supported<br />

H6: There is a positive relationship between top management support and implementation<br />

success Supported<br />

H7: There is a positive relationship between training and implementation success Supported<br />

H8: There is a positive relationship between vertical architecture and implementation success Supported<br />

6. Conclusion and future directions<br />

This paper presents several key findings as identified through a formal study, and improves the power of<br />

the explanatory success model, which extends and enhances prior literature and understanding of BI.<br />

The first key finding is the identified implementation construct addition of a vertical architecture,<br />

particularly for the small-medium business (SMB) scenario. A vertically-integrated architecture improves<br />

the implementation timeline and required resource base for implementation success. The second key<br />

finding is around establishing a collaborative culture to promote organizational learning capabilities. This<br />

extends previous notions of team work and business-IT alignment, and is necessary to support adoption<br />

and use of BI. Third involves implementation success outcomes when democratization or universal user<br />

adoption of BI has been achieved. In past studies measuring BI success, only a small portion of users<br />

had access to BI capabilities, successful outcomes can be realized while extending BI benefits to all<br />

users.<br />

Limitations include the use of single organization study; additional organizations should be reviewed to<br />

increase sample size in and across various industries and firm sizes. There is an importance to<br />

identifying universally applying critical success factors, and the ability for tailoring those factors to an<br />

individual organization or implementation. However it is the adaptability of the BI capabilities and the<br />

overall project that will ensure successful completion. Other identified areas of study and importance<br />

beyond implementation include establishing a competency center to ensure continue usage of business<br />

intelligence, stakeholder satisfaction, and decreased costs. Another area is establishing an architecture<br />

roadmap for future iterations and system updates, these include enhancements to key capabilities and<br />

features, bug fixes, security improvements, and ensured vendor support. A governance plan is also<br />

important to establish technical roles, support service level agreements, back and recovery, database<br />

and data standards, metadata standards, content branding, life cycle policies, and training.<br />

References<br />

Arnott, D. "Success Factors for Data Warehouse and Business Intelligence Systems," Australasian <strong>Conference</strong> on<br />

Information Systems) 2008, pp 55-65.<br />

Bernard, A. "Four Technology Best (and not so best) Bets for 2009," in: CIO Update, 2009.<br />

Bhatti, T.R. "Critical Success Factors for the Implementation of Enterprise Resource Planning (ERP): Empirical<br />

Validation," The Second International <strong>Conference</strong> on Innovation in Information Technology) 2005.<br />

Davenport, T.H., Harris, Jeanne G. Competing on Analytics Harvard Business School Publishing Corporation,<br />

Boston, 2007.<br />

Garson, G.D. "Partial Least Squares Regression (PLS)," N.C.S. University (ed.), 2010.<br />

Howson, C. Successful Business Intelligence: Secrets to Making BI a Killer App, 2008.<br />

Leidner, D., Elam, Joyce "Executive Information Systems: Their Impact on Executive Decision Making," Journal of<br />

Management Information Systems (101:3) 1993, pp 139-155.<br />

Moore, G.C., Benbasat, Izak "Development of an Instrument to Measure the Perceptions of Adopting an Information<br />

Technology Innovation," Information Systems Research (2:3) 1991, pp 192-222.<br />

Moss, L.T., Atre, Shaku Business Intelligence Roadmap, Boston, 2007.<br />

Rainer, R.K., Watson, Hugh "The Keys to Executive Information System Success," Journal of Management<br />

Information Systems (12:2) 1995, pp 83-98.<br />

561


Joseph Woodside<br />

Ranjan, J. "Business Justification with Business Intelligence," The Journal of Information and Knowledge<br />

Management Systems (38:4) 2008, pp 461-475.<br />

Ringle, C.M., Wende, S., Will, S. "SmartPLS 2.0 (M3) Beta," Hamburg, 2005.<br />

Saha, G.K. "Business Intelligence Computing Issues," ACM Ubiquity (8:25) 2007.<br />

Williams, S., Williams, Nancy The Profit Impact of Business Intelligence, 2007.<br />

Wixom, B.H., Watson, Hugh J. "An Empirical Investigation of the Factors Affecting Data Warehouseing Success,"<br />

MIS Quarterly (25:1) 2001, pp 17-41.<br />

562


Work<br />

in<br />

Progress<br />

Papers<br />

563


564


The Evolution of IT Governance in Multiple Industry Units<br />

and the Business Case for an Outside Partner<br />

Ganeshprasad Chandrasekaran 1, 2 , Chandramohan Annavarapu 2 and Balaji<br />

Rathinasamy 2<br />

1<br />

Tata Consultancy Service Limited, Chennai, India<br />

2<br />

SRM School of Management, SRM University, Chennai, India<br />

cgprasad@gmail.com<br />

profacmohan@gmail.com<br />

balaji99@gmail.com<br />

Abstract: IT Governance has become the integral part of the strategic thinking process for the growing number of<br />

companies across different industries. Companies are increasingly evaluating what is core to their business and<br />

weighing the benefits of outsourcing critical non-core functions such as IT (Information Technology). IT Governance<br />

exists within the context of corporate governance and the principles are especially the same. IT governance is an<br />

accountability framework and management process that helps to define and communicate what must be done and<br />

provides the rigorous oversight to ensure that it is. The accountability framework is typically planned with well defined<br />

roles and responsibilities reflecting decision rights among the stakeholders in the IT Management process and is<br />

reinforced with effective reporting. This paper reports on an investigation which focused on the role of IT governance<br />

and the outsourcing partner as mechanism for the effective and efficient management of IT. Results are presented<br />

from a survey in which respondents consisted of IT executives with responsibilities for IT governance and<br />

Outsourcing from different parts of the globe, who represent multiple industries. The outcome of the survey indicates<br />

a strong correlation between the overall use of IT governance practices and overall governance performance.<br />

Keywords: IT governance, outsourcing, corporate governance, business<br />

1. Introduction<br />

The advent of the 21 st century has brought about new challenges for the corporate world as it continues<br />

to evolve in adapting to ever-changing operating environments. However, business expectations for IT<br />

are rising. The vast majority of business processes are enabled by computers and organizations have no<br />

fallback paper processes. IT-enabled products and services have become a competitive differentiator in<br />

almost every industry. To help meet the strategic needs of the business, many companies are<br />

implementing IT governance practices to enhance IT and business alignment. IT governance is<br />

recognized as an extension of corporate governance. The IT Governance Institute (ITGI, 2008) indicates<br />

that corporate governance is the “Methodology by which a corporation is directed, administered and<br />

controlled, whereas IT governance supports achieving corporate objectives, strategy, direction,<br />

administration and control, using appropriate IT investment and resource management.” This paper<br />

reports on investigated results about “IT governance practices being used to balance the competing IT<br />

priorities like control, agility, cost, customer, business needs, revenue growth and profitability”.<br />

2. Literature review<br />

IT governance is the responsibility of executives and the board of directors. It consists of the leadership,<br />

organizational structures and processes that ensure that the enterprise’s IT sustains and extends the<br />

organization’s strategies and objectives (ITGI, 2008). Leveraging IT successfully to transform the<br />

enterprise and create value added products and services has become a universal business competency.<br />

IT is fundamental for managing enterprise resources, dealing with suppliers and customers, and enabling<br />

increasingly global transactions (Eric and William, 2008).<br />

According to the ITGI, “Boards and executive management generally expect their enterprise’s IT to<br />

deliver business value, i.e., provide fast, secured, high-quality solutions and services; generate<br />

reasonable return on investment; and move from efficiency and productivity gains toward value creation<br />

and business effectiveness (Alan, 2009). The governance of outsourcing is the function described as “it is<br />

no longer a company’s ownership of capabilities that matters but rather its ability to control and make the<br />

most of critical capabilities, whether or not they reside on the company’s balance sheet” (Lucin and<br />

Assaf, 2009). Outsourcing in simple terms is “the delegation of few services or IT-intensive business<br />

processes to an external provider for increasing the availability of service and the overall profitability”<br />

(Spremic, et al., 2008, p.252).<br />

565


3. Research findings<br />

Ganeshprasad Chandrasekaran et al.<br />

In addition to the literature review, the research study was supported by a survey and an interview. The<br />

survey was conducted with twenty one IT professionals and interview with three IT professionals. Both<br />

the survey and the interview were with IT professionals who fall under the category of senior executives<br />

and work for different CMM (Capability Maturity Model) organizations. During the interview the same<br />

survey questions were used, however had in detail discussion around every question. The literature<br />

review was instrumental in designing the survey questions and the survey has targeted use of the IT<br />

governance practices and its performance. The below points were used to frame the survey and they are<br />

(a).Are IT Governance practices able to balance the internally focused cost and control objectives?<br />

(b).Are IT Governance practices able to balance the externally focused customer and business<br />

needs?<br />

(c).Is IT Governance helps to meet the internal and external business objectives?<br />

(d).Can IT Governance practices ensures to prioritize the business demands?<br />

(e).Is there any real growth in revenue and profitability later to the implementation of IT governance?<br />

(f).What is the proportional focus of IT governance and an Overall Business Objective?<br />

(g).In General, are IT Governance initiatives working?<br />

The overall use of IT governance practices was measured by the average combined score and the<br />

results of the survey respondents indicates a strong correlation between the overall use of IT governance<br />

practices and overall governance performance. For the open-ended questions raised on IT governance,<br />

answers indicate that most common IT governance objectives are focused on cost containment (includes<br />

efficiency, standardization and automation) and risk reduction (includes compliance, security, and public<br />

scrutiny of IT failure). This is perhaps not surprising, given the recent economic climate and global surge<br />

in regulatory requirements imposed on IT organizations.<br />

Survey Response<br />

35<br />

30<br />

25<br />

20<br />

15<br />

10<br />

5<br />

0<br />

IT Goverance - by Maturity<br />

Cost Complaince Cusotomer Decision<br />

Category<br />

Figure 1: IT Governance objectives by maturity group<br />

High<br />

Medium<br />

Maturity is determined by a combination of use of governance practices and performance. The study<br />

revealed (as shown in Figure 1) the following about the different maturity groups.<br />

(i). Low maturity organizations (those with lower usage of IT governance practices and lower<br />

performance scores) have IT governance objectives that are more focused on cost and risk reduction.<br />

The high-impact IT governance practices that are shown to optimize performance at the initial level of<br />

maturity, however move beyond cost and risk to also focus on agility as well as project and service value<br />

that may not be cost oriented.<br />

(ii). Medium maturity organizations (those with moderate IT governance practice usage and medium<br />

performance scores) also focus on cost objectives, but they have a greater focus on customer and<br />

business related objectives. The high-impact IT governance practices at this maturity level build on the<br />

foundational practices to include supporting a process culture, asset management, tactical demand<br />

management and improving the transparency of project assumptions.<br />

566<br />

Low


Ganeshprasad Chandrasekaran et al.<br />

(iii). High maturity organizations (those with broadest use of IT governance practices, and highest<br />

performance scores) have the highest focus on customer and business related objectives and the lowest<br />

level of focus on cost containment. Mature organizations have learned to use IT auditing as a tool to<br />

achieve their operating objectives as well as addressing compliance mandates.<br />

4. Conclusion<br />

Although best practices are openly available and clearly described in literature, they are not necessarily<br />

widely adopted. This implies that in many organizations, there is significant room for improvement in the<br />

area of IT governance specifically in the low and medium matured organizations. As the IT organization<br />

makes the transition from utility service provider to a more strategic revenue enabler, IT governance can<br />

be a powerful tool for balancing competing IT priorities. However, this study suggests that in many<br />

organizations, IT governance initiatives are more focused on cost containment and risk reduction<br />

objectives. Focus on customer differentiation and enabling business needs has taken a lower priority than<br />

more foundational cost and risk reduction objectives. The results, however also indicate a positive<br />

element, IT governance initiatives are working and the further research will continue to understand the<br />

methods adopted by organizations to utilize the IT governance principles to the fullest extend.<br />

References<br />

Alan, C. (2009) IT Governance: Implementing Frameworks and Standards for the corporate Governance of IT, ITGI<br />

Publications, Illinois.<br />

Andreas, W.M. (2008) “Modes of Governance in Business Process Outsourcing: Equity Control or Tradability of<br />

Services, an Executive versus Market's Perspective”, Proceedings of the XLIst Annual Hawaii International<br />

<strong>Conference</strong> on System Sciences, Hawaii, USA.<br />

Brown, A.E. and Grant, G.G. (2005) “Framing the Frameworks: A review of IT Governance Research“,<br />

Communications of the Association for Information Systems, Vol. 15, pp 696-712.<br />

David, N. (2005) IT Governance: Managing Information Technology for Business, Thorogood Publications, London.<br />

Eric, J.B and William, A.Y (2009) The effective CIO: How to achieve outstanding success through strategic<br />

Alignment, Financial Management and IT Governance, Auerbach Publications, Florida.<br />

IT Governance Institute. (2008) Aligning COBIT®4.1, ITIL® V3 and ISO/IEC 27002 for Business Benefit. [online],<br />

Available at: < http://www.isaca.org/Knowledge-Center/Research/ResearchDeliverables/Pages/Aligning-COBIT-<br />

4-1-ITIL-V3-and-ISO-IEC-27002-for-BusinessBenefit.aspx > [Accessed on 10 November 2010].<br />

Lawrence, J.A, Susan, P., Gary, F.P and Dasaratha V.R (2007) “Corporate Governance, Audit Quality and the<br />

Sarbanes-Oxley Act: Evidence from Internal Audit Outsourcing“, The Journal of Accounting Review, Vol. 82, pp<br />

22-31.<br />

Lucian, A.B. and Assaf, H. (2009)”The Elusive Quest for Global Governance Standards“, University of Pennsylvania<br />

Law Review, Vol. 157, pp 1263-1317.<br />

Patel, N.V. (2002) “Global Ebusiness IT Governance: Radical Re-Directions”, Proceedings of the XXXVth Hawaii<br />

International conference on System Sciences, IEEE Computer Society, Hawaii, USA.<br />

Short, J. and Gerrard, M. (2009) “IT Governance Must be Driven by Corporate Governance” Gartner Research,<br />

Stamford, USA, November.<br />

Spremic, M., Zmirak, Z. and Kraljevic, K. (2008) “Evolving IT Governance Model – Research Study on Croatian<br />

Large Companies”, WSEAS Transactions on Business and Economics, Issue 5, Vol. 5, May, pp 250-259.<br />

Weill, P. and Ross, J.W. (2004) “IT Governance on One Page”, Center for Information Systems Research,<br />

Massachusetts Institute of Technology, Cambridge, USA.<br />

567


The Relationship Between Quality Management and<br />

Knowledge Management in the Service Industries<br />

Amir Honarpour and Ahmad Jusoh<br />

Management and Human Resource Development, University Technology<br />

Malaysia, Johor, Malaysia<br />

Amir_honarpour@yahoo.com<br />

ahmadj@utm.my<br />

Abstract: Regarding the knowledge based economy master plan (2002), Malaysia is transforming itself towards a<br />

knowledge-based economy, which concentrates on knowledge, people and virtual networks. Researchers argue that<br />

achieving such an objective and gaining value from knowledge causes some problems including changes in<br />

customer demands, developing new skills, knowledge creation, and modifying the economical structure to a<br />

knowledge-based economy etc. (Plump and Zamfir, 2009; Suete, 2001;Nonaka and Teece, 2001). These challenges<br />

affect the way organizations collect, use and disseminate information. Knowledge management can lead to a<br />

sustainable competitive advantage. On the one hand, it is supposed that knowledge is an intangible and unique<br />

asset, which provides an organization with competitive advantage, while on the other, QM contributes significantly to<br />

the performance of organizations, mostly in the manufacturing sector. Considering the organizational life and<br />

structure, both QM and KM are professional communities that are positioned at different points in their maturity<br />

lifecycle, however, they have similar aims and positions in regard to management. It appears that both quality<br />

management and knowledge management are complementary, if not compatible, as the strategies of both are long<br />

term for the intention of gaining competitive advantage and overall improved performance. QM aims to help the<br />

value-creators of organizations in improving their performance as well as KM. They hold somesimilar basic<br />

assumptions, e.g. the importance of cultural changes and process improvement. There are valuable studies on the<br />

description of the interaction between QM and KM (see Jaime et al., 2006; Dvir, 2004). Regarding this point of view,<br />

this research is going to consider implementing KM in an organization with regard to QM practice and show how they<br />

interact and how they improve their effectiveness.<br />

Keywords: knowledge, knowledge management, quality management<br />

1. Introduction<br />

In the early 1990s, Malaysiaset out a 30-year plan as vision 2020 to become a developed country in<br />

terms of economic performance and technological capability. Transferring from a manufacturing industrial<br />

economy to a knowledge-based economy is a part of this plan (Ramlee and Abu, 2005; Fleming and<br />

Soborg, 2010). A knowledge-based economy (k-economy) focuses on knowledge, people and virtual<br />

networks, and requires the adoption and application of new management and organizational systems<br />

along with knowledgeable, skilled, dynamic, creative and innovative human resources (Plumb and Zamfir,<br />

2009; Ramlee and Abu, 2005; Fleming and Soborg, 2010). Moving towards a knowledge-based economy<br />

needs intensive knowledge investment such as research & development (R&D), software, information<br />

and communication technology (ICT), education and training (Suete, 1999). To succeed in this way,<br />

knowledge management has an essential role. As Chong et al. (2010) suggested, knowledge<br />

management (KM) has a strong effect on innovation adoption. In addition, to support scientific activities,<br />

especially in R&D, uses of quality management complemented by knowledge management will improve<br />

the quality of research (Jaime, 2006; Holt and Dilani 2009). Finally, the object of education and training<br />

should be to make quality human resources and knowledge workers the backbone of the k-economy.<br />

Such a labour force is able to create and affect knowledge to produce useful actors with strong and<br />

analytical skills (Ramlee and Abu, 2005). According to Abadesco (2004), a knowledge worker is ‘anyone<br />

who makes a living out of creating, manipulating or disseminating knowledge’. Therefore, KM is the<br />

infrastructure for knowledge workers' jobs, and also has an important role in supporting activities required<br />

for achieving a K-economy. However, the point is that recent studies have revealed that a large amount<br />

of KM projects (more than 50%) fail or do not achieve their goals (Prusak and Weiss, 2007;Madanmohan,<br />

2005). This may be caused by many reasons such as the lack of top management commitment or failing<br />

to locate, capture, organize and disseminate the tacit knowledge (see Akhavanet al., 2005).<br />

Researchers have linked Quality Management (QM) with organizational learning in an attempt to show<br />

how Quality Management leads to continuous improvement (Choo et al., 2007; Konidari and Abernot,<br />

2006). Linderman et al. (2004) stated that knowledge is created by quality management practices as<br />

shown in the Nonaka's Knowledge Creation Process. Molina et al. (2007) advocated that QM practices<br />

have an important role in the transferring of knowledge. In addition, the knowledge-based economy is<br />

replacing the meaning of service quality management in a way that it becomes knowledge-driven and<br />

568


Amir Honarpour and Ahmad Jusoh<br />

relies on people’s continuous growth and sharing ideas and knowledge (Plumb and Zamfir, 2009). It<br />

seems that the notion of knowledge and quality are strongly associated, as Jaime(2006) said "as<br />

Knowledge Management reaches its maturity in terms of acceptance as an important part of doing<br />

business in the modern world, quality will again become the mantra of successful companies".<br />

Based on the literature, quality management is a good basis and facilitator for implementing knowledge<br />

management, however, empirical research on this subject is insufficient. In Malaysia 2,383 organizations<br />

have received ISO 9001:2000 and other firms are mostly ready to execute IMS in their organization<br />

(Kadir et al., 2009). As QM practices and systems are widely used in the country, this research can be<br />

helpful concerning Vision 2020.<br />

2. Knowledge management and quality management<br />

QM has contributed significantly to the performance of organizations, mostly in the manufacturing sector.<br />

It played a major role in the rise of the Japanese industry in the 60's and 70's (James, 1997), and the<br />

recovery of some sectors of the American industry in the 80's. In the 80's and early 90's there was<br />

growing management attention on QM, which was accompanied by a considerable growth in the<br />

community. There are different QM models and techniques such as ISO 9000, Deming Prize, the<br />

American Malcolm Baldrige National Quality Award, and the <strong>European</strong> Quality Award. Many researchers<br />

from the mid 1990s have studiedthe field of knowledge management and have contributed to relevant<br />

theories. In this study, the systematic framework of organizational knowledge management processes<br />

introduced by Alavi and Leidner (2001) is applied. This framework is based on the sociology of<br />

knowledge and considers the cognitive and social nature of organizational knowledge. It also represents<br />

an individual's cognition and practices as well as the organizational ones (Alavi and Leidner, 2001). The<br />

knowledge management processes include knowledge creation, knowledge storage/retrieval, knowledge<br />

transfer, and knowledge application. According to Pentland (in Alavi and Leidner, 2001), knowledge<br />

creation refers to developing new knowledge or replacing existing content within the organization's tacit<br />

and explicit knowledge. Knowledge storage or retrieval refers to the knowledge that resides in different<br />

component forms, such as written documentation or structured information stored in electronic<br />

databases. Knowledge transfer refers to transferring knowledge between individuals or from individuals to<br />

organizational memory or across groups. Knowledge application is defined as the integration of<br />

knowledge for creating organizational capability (Alavi and Leidner, 2001).<br />

Regarding to previous studies the conceptual model has been developed. This model is introduced in<br />

figure 1. Based on this model some propositions has been presented to investigate the relationship<br />

between QM and KM.<br />

Figure 1: Conceptual framework of the study<br />

569


2.1 QM and knowledge creation<br />

Amir Honarpour and Ahmad Jusoh<br />

The application of information technology and systems has recently attracted traditional service providers<br />

such as airlines and travel agencies into service-value networks (Plumb and Zamfir, 2009). These<br />

changes are towards the satisfaction of the changing needs of customers. Customer satisfaction also<br />

means considering internal customers (Yusof and Eng, 2003). This relationship can be provide by<br />

listening mechanisms, QFD, customer stories etc. Linderman (2004) claimed that, "Quality management<br />

practices that promote monitoring and feedback of customer information allow knowledge to be created".<br />

Practices such as customer relationship and workforce management. Therefore, we propose the<br />

following proposition:<br />

Proposition1: Quality management practices that are related to customer satisfaction can affect the<br />

knowledge creation process.<br />

2.2 QM and knowledge storage/retrieval<br />

One of the quality management values is management by facts. This means that all the decisions should<br />

be based on documented facts (Lagrosen and Lagrosen, 2006). This concept relies on practices and<br />

tools that are used in different QM practices and techniques such as quality information, the idea bank,<br />

flowcharts, and decision making tools. This value is a base for documenting culture and is considered as<br />

a facilitator for the use of organizational memory, such as databases and repositories. Hence, the<br />

following proposition is formulated:<br />

Proposition2: Quality management practices affect the knowledge storage/retrieval process.<br />

2.3 QM and knowledge transfer<br />

One of the principals of quality management is establishing teams (Molina et al., 2006). According to<br />

Linderman et al. (2004), teams begin to develop ideas to resolve quality problems. These team meetings<br />

are a place in which knowledge is transferred. There are also other practices such as process<br />

management and product/service design that use some specific tools like SPC, cause and affect analysis<br />

and statistical charts, which lead to transfer knowledge from a separate part of the organization to other<br />

parts. Training and education are also important factors of QM. Such practices and tools transfer<br />

knowledge from individuals to individuals (e.g. team meetings in process management practice) or from<br />

individuals to group (e.g. educations in workforce management). Thus, the following proposition could be<br />

formulated:<br />

Proposition3: Quality management practices and tools affect the knowledge transfer process<br />

2.4 QM and knowledge application<br />

To apply knowledge management, Demsetz (in Alavi and Leidner, 2001) identified three steps of the<br />

knowledge integration process: first, directives or a set of rules, standards, and procedures; second,<br />

organizational routines or interaction protocols, and process specification; and third, self-contained task<br />

teams or teams of individuals with prerequisite knowledge for problem solving (Alavi and Leidner, 2001).<br />

As mentioned earlier, teams facilitate developing new ideas to resolve quality problems (see Linderman<br />

et al., 2004). System flowcharts and process flowcharts are QM tools, which identify what scope of<br />

information is required to do work or what codes and standards govern the work (Shearer, 1994). Also<br />

the main aim of process management practice is to reduce process variation to reduce reworks and<br />

make stable process. Therefore, this study proposes that:<br />

Proposition4: Quality management practices and tools affect the knowledge application process<br />

2.5 Senior management commitment<br />

Senior management commitment is a necessary condition for the implementation of quality management<br />

(Gitlow, 2001). This commitment is one of the values of quality management implementation and, as<br />

Lagrosen and Lagrosen (2006) said, "leaders can excel if only their hearts are into their quality journey<br />

and they have a determination to carry it through". Quality management is a long-term strategy for<br />

gaining competitive advantage and improving overall performance. It aims to improve quality and<br />

productivity. Therefore, senior managers who are committed to improve quality and productivity through<br />

QM are more likely to invest in knowledge management. As a result:<br />

570


Amir Honarpour and Ahmad Jusoh<br />

Proposition 5: Top management commitment to QM will affect the acceptance of implementing KM<br />

3. Conclusion<br />

In conclusion, the implementation of knowledge management can be conceptualized as quality<br />

management practices. The conceptual model proposes that the commitment of senior managers, quality<br />

management practices which promote monitoring and feedback of customer information, management by<br />

facts, teamwork, problem solving tools and process orientation work can improve the success of<br />

knowledge management implementation. The next step in this investigation is to develop a survey to<br />

examine the model and its proposed relationships.<br />

References<br />

Akhavan, P., Jafari, M. and Fathian, M. (2005) Exploring Failure-Factors Of Implementing Knowledge Management<br />

Systems In Organizations, Journal of Knowledge Management Practice<br />

Alavi, M. Leidner, D. (2001) Knowledge Management and Knowledge Management System: Conceptual Foundations<br />

and Research Issues, MIS Quarterly, Vol. 25, No. 1, pp: 107-136<br />

Abadesco, V. (2004) Training knowledge worker,Asian Productivity Organization, Tokyo<br />

Chong, A. Y., Ooib, K., Linc, B. and Tehb, P. (2010) TQM, knowledge management and collaborative commerce<br />

adoption: A literature review and research framework, Total Quality Management, Vol 21, No. 5, pp: 457–473<br />

Choo, A. S., Linderman, K. W. and Schroeder, R. G. (2007) Method and context perspectives on learning and<br />

knowledge creation in quality management, Journal of Operations Management, Vol 25, pp: 918–931<br />

Dvir, R. (2000) Quality By Knowledge: Collective Creation of A New Concept, Quality Progress, August 2000<br />

Eng, Q, E., Yusof, S, M. (2003) A survey of TQM practices in the Malaysian electrical and electronic industry, Total<br />

Quality Management, Vol 14, No 1, pp: 63-67<br />

Fleming, D. and Soborg, H. (2010) Malaysia’s Human Resource Strategies for a Knowledge-Based Economy –<br />

Comparing the Influence of Different Labour Market Relations, <strong>European</strong> Journal of Social Sciences, Vo 16,<br />

Number 2, pp:286-306<br />

Gitlow, H. S. (2001) Quality management systems: a practical guide, St. Lucie Press<br />

Holt, R. and Dilani, J. (2009) Knowledge and quality management: An R&D perspective, Technovation, Vol29, pp:<br />

775–785<br />

Jaime, A., Gardoni, M., Mosca, J. and Vink, D. (2006) From Quality Management to Knowledge Managementin<br />

Research Organisations, IJIM, Vol10, No. 2, pp:197-215<br />

James, D. (1997). TQM Seeks a New Relevancy in the Emerging 'Knowledge Economy', Business Review Weekly<br />

(Australia),September22, p 199<br />

Kadir, A. et al (2009) Implementation of Integrated Management System in Malaysia: The Level of Organization’s<br />

Understanding and Awareness, <strong>European</strong> Journal of Scientific Research, vol 31, No. 2, pp.188-195<br />

Konidari, V. Abernot, Y (2007) From TQM to learning organization: Another way for quality management in<br />

educational institutions, International Journal of Quality & Reliability Management, Vol 23, No. 1, pp: 8-26<br />

Lagrosen, S. and Lagrosen, Y. (2006) A dive into the depths of quality management, <strong>European</strong> Business Review,<br />

V18, N.2, pp: 84-96<br />

Linderman, K., Schroeder, R. G., Zaheer, S., Liedtke, C. and Choo, A. S. (2007) Integrating quality management<br />

practices with knowledge creation processes, Journal of Operations Management, Vol22, pp: 589–607<br />

Madanmohan, R. (2005) Knowledge management tools and techniques: practitioners and experts, Elsevier, Oxford<br />

Molina, L. M., Lloréns-Montes, J. and Ruiz-Moreno, A. (2007) Relationship between quality management practices<br />

and knowledge transfer, Journal of Operations Management, Vol25, pp: 682–701<br />

Nonaka, I. and Teece, D. (2001). Managing Industrial Knowledge: Creation, Transfer and Utilization, SAGE<br />

publications, London<br />

Ramlee, M. and Abu, A. (2005) Malaysia transitions toward a knowledge-basedEconomy, The Journal of Technology<br />

Studies, 51-61<br />

Plumb, I. and Zamfir, A. (2009) Managing service quality within the knowledge based economy: opportunities and<br />

challenges, Quality Management in Services, VolXI, No 26<br />

Prusak, L. and Weiss, L. (2007) Knowledge Creation and Management: New Challenges for Managers, Oxford<br />

University Press<br />

Shearer, C. (1995) Practical Continuous Improvement for Professional Services, American Society for Industrial<br />

Security<br />

571


Construction and Validation of eSchool Success Model<br />

Hesbon Nyagowa 1 , Dennis Ocholla 1 and Stephen Mutula 2<br />

1 University of Zululand, South Africa<br />

2 University of Botswana, Botswana<br />

nyagowa_hesbon@yahoo.co.uk<br />

docholla@pan.uzulu.ac.za<br />

MUTULASM@mopipi.ub.bw<br />

Abstract: DeLone and McLean (1992) IS Success Model has been widely adapted in evaluating success of many<br />

different information systems including e-commerce, e-learning and knowledge management systems. Adapting the<br />

D&M IS success model requires re-specification to suit specific systems. The re-specified models tend to introduce<br />

new dimensions and therefore new measuring instruments that call for empirical validation. In the current study<br />

evaluating NEPAD eSchool pilot programme in Kenya, a distinct new dimension was added to the D&M IS success<br />

model and the net-benefit dimension measure was re-defined to capture the objectives of NEPAD eSchool<br />

programme. The study specified seven dimensions to eSchool system: infrastructure quality, information content and<br />

communication quality, technical personnel service quality, training effectiveness, extent and exploitative use, user<br />

satisfaction and net-benefits. The eSchool success model was conceptualised theoretically and subsequently<br />

validated empirically using data collected from 776 students and teachers in six schools which have implemented<br />

NEPAD eSchool system. The data was split into two to facilitate implementation of both exploratory factor analysis<br />

(EFA) and confirmatory factor analysis (CFA). The first half of the data set was used in performing principal<br />

component analysis on one new dimension (training effectiveness) and the re-specified dimension (net-benefit). The<br />

second half of the data set was used in confirmatory factor analysis to test the goodness-of-fit of the eSchool<br />

success model in evaluating eSchool success. This paper discusses the eSchool success model construction, its<br />

validation outcome and the contribution of the study to IS evaluation. This is part of a study in progress.<br />

Keywords: eSchool, eSchool success model, validation of eSchool system, validation, IS evaluation<br />

1. Introduction<br />

1.1 eSchool as an Information System (IS)<br />

In its original patent, the eSchool teaching system has four sub-systems; students’ sub-system, teachers’<br />

sub-system, library sub-system and administration sub-systems (Nobles et al., 1989) all interfaced such<br />

that user groups are restricted to modules and content relevant for their roles. Information system<br />

consists of computers, communication networks, instructions (application modules), stored files, people<br />

and procedures. ISs are characterized in organizational work system (Alter, 2008) and serve as<br />

information management tools. They therefore enable user groups to engage in retrieval of data, sharing<br />

of data, manipulation of information, and transmission of information. ESchool system enables students<br />

and teachers to share information through communication network inbuilt in the computer systems,<br />

facilitate storage, sharing and retrieval of data by users and user groups.<br />

Specifically, eSchool systems are designed to avail content relevant to curriculum, sharing of class work<br />

materials between teachers and students, students & students, and teachers & teachers’, retrieval of<br />

curriculum support materials from other data bases, communication between users, and management of<br />

student data within the school network. Teaching, learning and education management is therefore<br />

accorded an alternative platform in eSchool system. These characteristics of eSchool systems qualify<br />

them as information system as defined in Alter (2008).<br />

1.2 Justification of need for eSchool evaluation<br />

Implementing eSchool system has three major cost components; infrastructure development, content<br />

development and upgrading of ICT expertise of students and teachers. In addition to ICT expertise<br />

upgrading, teachers require retraining on pedagogical approaches that reduce their direct involvement in<br />

the learning process (Hennessy, Ruthven & Brindley, 2005) as sage-on-the-stage to facilitators. ESchool<br />

infrastructure, content and human skills development require high capital outlay that needs<br />

rationalization. To ascertain sustainability and success of an information system, evaluation mechanism<br />

should be put in place to give indicators of its success and re-strategizing the implementation to improve<br />

the success rate.<br />

572


Hesbon Nyagowa et al.<br />

Traditionally, IS programmes have had evaluation inbuilt as part of IS development cycle (Gremberen,<br />

2001). The theories and practices behind the traditional evaluations have tended to be specific to the<br />

programme and they deploy varied non-standardized tools which deny IS community opportunity to<br />

compare evaluation outcomes (DeLone & McLean, 1992; Petter, DeLone & McLean, 2008). Studies<br />

which have used similar evaluation models such as DeLone and McLean (1992) IS success model have<br />

concentrated on one or two dimensions of the model. Absence of methodological comparability limits the<br />

trend in knowledge growth particularly in extension and generalization of knowledge (Petter, DeLone and<br />

McLean, 2008) in new environment. It is in this respect that Petter, DeLone and McLean (2008) urged IS<br />

success evaluation researchers to use validated instruments in measuring success of any IS.<br />

Proliferation of eSchool systems has been experienced in many countries in the recent past (Simpson,<br />

Payne & Condie, 2005; Condie & Livingston, 2007; Ya’Acob, Nor & Azman, 2005; UNICT Task Force,<br />

2005). The force behind this trend is the high expectations organizations have placed on the potentials of<br />

eSchool system; studies converge on the possible benefits of eSchools which include productivity<br />

improvement in the education sector, imparting lifelong learning skills to students and creating team work<br />

skills amongst students through collaboration between different learning communities (EL-Halawany &<br />

Huwail, 2008). To safeguard against moribund projects, eSchool researchers should seek dependable<br />

success evaluation model and instruments.<br />

The current study conceptualized an eSchool success evaluation model based on DeLone and McLean<br />

(1992) IS success model. The theoretically conceptualized model was empirically validated to establish<br />

consistency and stability using factor analysis. This poster presents the conceptualized model and the<br />

results of its validation.<br />

2. Methods<br />

2.1 Conceptualization of eSchool success model<br />

Following analysis of several IS success evaluation studies, DeLone and McLean (1992) developed an IS<br />

success model which has seven dimensions namely; system quality, information quality, intention to use,<br />

use, user satisfaction, individual impact and organization impact (DeLone & McLean, 1992). The model<br />

has been validated (ICIS, 2007; Wu & Wang, 2006; Bokhari, 2005; Seddon & Kiew, 1994; Roldan & Leal,<br />

2003) and augmented (Pitt, Watson & Kavan, 1995; DeLone & McLean, 2003; Adeyinka, 2009; DeLone<br />

& McLean, 2008). In its present form D&M IS success model’s seven dimensions are schematically<br />

represented in figure 1.<br />

Information<br />

Quality<br />

System Quality<br />

Service Quality<br />

Intention to Use<br />

Use<br />

User Satisfaction<br />

Net Benefit<br />

Figure 1: D&M IS success model’s seven dimensions (source: DeLone and McLean (2003))<br />

573


Hesbon Nyagowa et al.<br />

DeLone and McLean IS success model (2003) has widely been accepted in many evaluation studies<br />

(Molla & Licker, 2001; Ong, Chorng-Shyong & Lai, Jung-Yu, 2004; Adeyinka, 2009) but each study has<br />

had to re-specify the model to suit the system being studied. Underpinning on extant literature, the<br />

current study conceptualized a seven dimensional eSchool success model. ESchool users (students and<br />

teachers) are introduced to eSchool use when they have very limited exposure to ICT literacy. Teachers<br />

in particular, were previously trained on pedagogies which had not anticipated integration of ICT in<br />

teaching and learning. Invariably, retraining of teachers on pedagogies consistent with eSchool<br />

requirements and imparting to both teachers and students ICT skills give them positive attitude to<br />

eSchool and may play a big role in the success of eSchool. The current study therefore conceptualized<br />

training effectiveness as a necessary dimension to eSchool success model. As training impacts on<br />

attitude, and intention to use IS is an attitudinal construct, the current study dropped the dimension<br />

‘intention to use’ in the D&M IS success model. Further, in eSchool, availability of content relevant to<br />

curriculum will to a great extent influence its adoption and use in teaching and learning. The current study<br />

therefore conceptualized that content should be given prominence more so for the fact that sustainability<br />

of eSchool partly rely on availability of content. Lastly, the study argues that accruing measurable<br />

benefits from eSchool may on its own drive more investment in eSchool thus causal relationship between<br />

‘net-benefit and eSchool infrastructure quality. The conceptualized eSchool success model is<br />

represented in figure 2.<br />

E-school<br />

infrastructure Quality<br />

Content and<br />

Communication<br />

Quality<br />

Service Quality of<br />

Technical<br />

Department<br />

Training for Use<br />

Effectiveness<br />

Figure 2: Modified D and M model for eSchools<br />

2.2 Validation design<br />

Use<br />

Teacher and Learners<br />

Satisfaction<br />

Ha<br />

Net Benefit<br />

The current study was implemented using NEPAD eSchool project pilot in Kenya as a case. NEPAD e-<br />

Africa Commission is piloting eSchool system in 17 African countries including Kenya. The eSchools<br />

have computers with local area network (LAN) and wide area network (WAN) established through V-SAT<br />

connection to the internet. The project has training component for users. The main objectives of NEPAD<br />

eSchool include improving the quality of graduating students. This would be evident in the extent to which<br />

students are better prepared to fit well in the modern work life. The modern work life is intensive in<br />

computer use and therefore students should be enabled to acquire computer skills, develop team work<br />

skills through collaborations and lifelong learning skills. Developing lifelong learning skills require learners<br />

to engage in self directed learning which is achieved through practicing acquisition of learning materials<br />

independently.<br />

574


Hesbon Nyagowa et al.<br />

The populations for the study were the six secondary schools:Chevakali High school in Western<br />

Province; Maranda High School in Nyanza Province; Menengai Mixed Secondary School in the Rift<br />

Valley Province; Isiolo Girls Secondary School in Eastern region; Mumbi Girls Secondary School in<br />

Central Province; and Wajir Girls Secondary School in North Eastern region. The other part of the<br />

population was users of the eSchool systems in the six schools. The six schools altogether had a<br />

population of 4,936 (1,887 F and 3,049 M) students with 250 teachers. A sample of 1,508 users was<br />

identified using stratified random techniques. Stratification was based on school population, class size in<br />

the different levels (years 9, 10, 11 and 12 of education) and gender (for co-education schools). A<br />

response rate of 51.4% was realized.<br />

Data collected was cleaned and entered into SPSS programme. Reliability and validity of the dater was<br />

assessed. Reliability was assessed using Cronbach Alpha while validity was tested using factor analysis<br />

applying principal component analysis.<br />

3. Results<br />

3.1 Reliability<br />

Test alpha obtained from first half of the data was found to be very closely associated with confirmatory<br />

alpha obtained from second half of the data as shown in table 1 and figure 3.<br />

Table 1: Test and confirmatory alpha<br />

Dimension Test Alpha Confirmatory Alpha<br />

Infrastructure Quality 0.6592 0.704<br />

Information Content and Communication Quality 0.7448 0.7416<br />

Service Quality 0.7182 0.754<br />

Training Effectiveness 0.3962 0.4818<br />

Extent and Exploitative Use 0.7893 0.7766<br />

User Satisfaction 0.916 0.9071<br />

Figure 3: Scatter plots of test and confirmatory alpha<br />

Net Benefits 0.8519 0.822<br />

575


3.2 Validity<br />

Hesbon Nyagowa et al.<br />

Scree plot revealed an elbow at a point lying between ordinate 5 and 9 as in figure 4. This confirmed<br />

existence of at least seven components to the eSchool success model as conceptualized.<br />

Eigenvalue<br />

Figure 4: Scree plot<br />

16<br />

14<br />

12<br />

10<br />

8<br />

6<br />

4<br />

2<br />

0<br />

1<br />

Scree Plot<br />

5<br />

9<br />

13<br />

17<br />

21<br />

Component Number<br />

25<br />

29<br />

3.3 Construct and factor loading<br />

Extraction table was drawn to identify the measurement items which fall under same factor. The mean<br />

factor loading was found to range between 0.4786 and 0.6625 as shown in table 2 and which is<br />

considered acceptable for exploratory study.<br />

Table 2: Mean factor loading<br />

Factor Mean Alpha<br />

Infrastructure Quality 0.5586<br />

Information Content and Communication Quality 0.6176<br />

Service Quality 0.5687<br />

Training Effectiveness 0.4787<br />

Extent and Exploitative Use 0.5534<br />

User Satisfaction 0.6835<br />

Net Benefits 0.6625<br />

4. Conclusion and recommendation<br />

4.1 Conclusion<br />

Confirmatory alpha for the data was generally high above 0.65 with the exception of ‘training<br />

effectiveness which is a new dimension in IS success under exploration. Scree plot confirmed existence<br />

of at least seven dimensions to the eSchool success model conceptualized. The item factor loading to the<br />

dimensions had a mean above 0.4 which is considered good construct validity for sample size of 200<br />

(Hair et al., 1998). Agreement between theoretical conceptualization and empirical validation is therefore<br />

strong enough to support suitability of the eSchool success model for evaluating other eSchool systems.<br />

576<br />

33<br />

37<br />

41<br />

45<br />

49<br />

53<br />

57<br />

61


4.2 Recommendation<br />

Hesbon Nyagowa et al.<br />

Of the seven dimensions, the training effectiveness achieved lowest item mean factor loading of 0.4787.<br />

Some items loading to this dimension were much lower. Since training effectiveness measurement<br />

instrument was exploratory, it would benefit the IS evaluation community to contribute towards refinement<br />

of the construct and its measurement. The study could not collect data from students who have<br />

completed schooling in eSchools and therefore could not measure the all the anticipated benefits of<br />

eSchool. It is recommended that future studied include graduates of eSchools to shed more lights on the<br />

benefits of eSchool. Lastly, with the good agreement between conceptual model and empirical testing,<br />

more studies should consider adopting the eSchool success model to determine its validity.<br />

References<br />

Adeyinka, T. (2009). An evaluation of WEBCT course content management system at the University of Botswana<br />

(Unpublished thesis).<br />

Condie, R. and Livingston, K. (2007). Blending online learning with traditional approaches: Changing practices,<br />

British Journal of Educational Technology, Vol. 38, No. 2, pp. 337-348.<br />

DeLone, H.W. and McLean, R.E. (2002). Information System Success Revisited, Proceedings of the 35 th Hawaii<br />

International <strong>Conference</strong> on System Science.<br />

Delone, W.H. & McLean, E.R. (1992) “Information systems success: The quest for the dependent variable”.<br />

Information Systems Research. 3(1), 60-95.<br />

EL-Halawany, H. and Huwail, E. I. (2008). Malaysian Smart Schools: A Fruitful Case Study for Analysis to Synopsize<br />

Lessons Applicable to the Egyptian Context, International Journal of Education Using ICT, Vol. 4, No. 2.<br />

Grembergen, W.V. (2001). Information technology evaluation methods & management, IDEA, Group Publishing,<br />

London, UK.<br />

Hair, J.F. jr., Anderson, R.E>, Tatham, R.C., and Black, W.C. (1998). Multivariate data analysis, 6 ed. Englewood<br />

Cliffs: Prentice Hall.<br />

International <strong>Conference</strong> on information System (2007). Using Market Exchange Theory to Conceptualize IS Quality<br />

and Re-Conceptualize the IS Success Model, Montreal.<br />

Molla, A. and Licker, S.P. (2001). E-commerce system success: An attempt to extend and respecify the DeLone and<br />

McLean model of IS success, Journal of Electronic Commerce Research, Vol. 2, No. 4, pp. 131-141.<br />

Nobles, A.A.et al. (1989). Electronic school teaching system, United States Patent Number 4820167.<br />

Ong, Chorng-Shyong and Lai, Jung-Yu (2004). Developing an instrument for measuring user satisfaction with<br />

knowledge management systems, proceeding of the 37 th Hawaii International <strong>Conference</strong> on System Science<br />

Petter, S., DeLone, W. and McLean, E. (2008). Measuring Information Systems Success: Models, Dimensions,<br />

Measures, and Interrelationships, <strong>European</strong> Journal of Information Systems, Volume 17, July 2008, pp. 236-<br />

263.<br />

Pitt, F.L., Watson, T.R., and Kavan, B.C. (1995). Service Quality: A measure of Information System Effectiveness,<br />

MIS Quarterly, Vol. 19, No. 2, pp. 173-187.<br />

Pitt, F.L., Watson, T.R., and Kavan, B.C. (1995). Service Quality: A measure of Information System Effectiveness,<br />

MIS Quarterly, Vol. 19, No. 2, pp. 173-187.<br />

Roldan, J.L., Leal, A. (2003). A Validation Test of an Adaptation of the DeLone and McLean’s Model in the Spanish<br />

EIS Field, Idea Group Publishing.<br />

Seddon, B.P. and Kiew, M. (1994). A partial test and development of DeLone and McLean’s Model of IS, Australian<br />

Journal of Information System, Vol.4, No. 1.<br />

Simpson, M., Payne, F. and Condie, R. (2005). Introducing ICT in Secondary Schools: A Context for Reflection on<br />

Management and Professional Norms, Education Management Leadership, 33; 331<br />

UNICT Taskforce (2005). Harnessing the Potential of ICT for Education: A Multi Stakeholder Approach, proceedings<br />

from the Dublin Global Forum for the United Nations ICT Task Force<br />

Wu, J-H. and Wang, Y-M. (2006). Measuring KMS Success: A respecification of DeLone and McLean’s Model,<br />

Information and Management, 43, 728-739.<br />

Ya’Acob, A., Nor, M.F.N and Azman H. (2005). Implementation of Malaysian Smart schools: An investigation of<br />

teaching-learning practices and teacher-student readiness, Internet Journal of e-Language Learning and<br />

Teaching, Vol. 2, No. 2, pp. 16-26.<br />

577

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!